2025-12-04T09:19:37.1528059Z Current runner version: '2.330.0' 2025-12-04T09:19:37.1533079Z Runner name: 'i-0c3ca63174fa64465' 2025-12-04T09:19:37.1533717Z Runner group name: 'default' 2025-12-04T09:19:37.1534395Z Machine name: 'ip-10-0-45-203' 2025-12-04T09:19:37.1536381Z ##[group]GITHUB_TOKEN Permissions 2025-12-04T09:19:37.1538300Z Contents: read 2025-12-04T09:19:37.1538782Z Metadata: read 2025-12-04T09:19:37.1539181Z ##[endgroup] 2025-12-04T09:19:37.1540781Z Secret source: Actions 2025-12-04T09:19:37.1541358Z Prepare workflow directory 2025-12-04T09:19:37.1919475Z Prepare all required actions 2025-12-04T09:19:37.1951969Z Getting action download info 2025-12-04T09:19:37.4865189Z Download action repository 'pytorch/test-infra@main' (SHA:39aa74d619174326f4e2fb0e216151c2f29d9ffd) 2025-12-04T09:19:40.4193032Z Download action repository 'pytorch/pytorch@main' (SHA:7716da9fb23f27a65b41f9f016a2afadf281c18f) 2025-12-04T09:19:57.5548267Z Download action repository 'actions/setup-python@a26af69be951a213d495a4c3e4e4022e16d87065' (SHA:a26af69be951a213d495a4c3e4e4022e16d87065) 2025-12-04T09:19:57.9293700Z Download action repository 'aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722' (SHA:ececac1a45f3b08a01d2dd070d28d111c5fe6722) 2025-12-04T09:19:58.2268197Z Download action repository 'aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076' (SHA:062b18b96a7aff071d4dc91bc00c4c1a7945b076) 2025-12-04T09:19:58.4523134Z Download action repository 'seemethere/download-artifact-s3@1da556a7aa0a088e3153970611f6c432d58e80e6' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-12-04T09:19:58.7313836Z Download action repository 'seemethere/upload-artifact-s3@baba72d0712b404f646cebe0730933554ebce96a' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-12-04T09:19:59.0251851Z Getting action download info 2025-12-04T09:19:59.1587818Z Download action repository 'actions/checkout@v4' (SHA:34e114876b0b11c390a56381ad16ebd13914f8d5) 2025-12-04T09:19:59.4381356Z Getting action download info 2025-12-04T09:19:59.5651888Z Download action repository 'nick-fields/retry@v3.0.0' (SHA:7152eba30c6575329ac0576536151aca5a72780e) 2025-12-04T09:19:59.7533481Z Getting action download info 2025-12-04T09:19:59.9087681Z Download action repository 'nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482' (SHA:3e91a01664abd3c5cd539100d10d33b9c5b68482) 2025-12-04T09:20:00.1644544Z Getting action download info 2025-12-04T09:20:00.3598244Z Uses: pytorch/pytorch/.github/workflows/_linux-test.yml@refs/heads/main (ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32) 2025-12-04T09:20:00.3601413Z ##[group] Inputs 2025-12-04T09:20:00.3601895Z build-environment: linux-jammy-py3.10-gcc11-build 2025-12-04T09:20:00.3608277Z test-matrix: {"include": [{"config": "cpu_inductor_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_avx2_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_timm", "shard": 1, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_timm", "shard": 2, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}]} 2025-12-04T09:20:00.3614674Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:20:00.3615272Z sync-tag: 2025-12-04T09:20:00.3616035Z timeout-minutes: 240 2025-12-04T09:20:00.3616312Z use-gha: 2025-12-04T09:20:00.3616534Z dashboard-tag: 2025-12-04T09:20:00.3616795Z s3-bucket: gha-artifacts 2025-12-04T09:20:00.3617075Z aws-role-to-assume: 2025-12-04T09:20:00.3617560Z disable-monitor: false 2025-12-04T09:20:00.3617815Z monitor-log-interval: 5 2025-12-04T09:20:00.3618283Z monitor-data-collect-interval: 1 2025-12-04T09:20:00.3618589Z ##[endgroup] 2025-12-04T09:20:00.3619013Z Complete job name: periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx) 2025-12-04T09:20:00.4077289Z A job started hook has been configured by the self-hosted runner administrator 2025-12-04T09:20:00.4153421Z ##[group]Run '/home/ec2-user/runner-scripts/before_job.sh' 2025-12-04T09:20:00.4160296Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:20:00.4160716Z ##[endgroup] 2025-12-04T09:20:01.5086901Z Runner Type: linux.8xlarge.amx 2025-12-04T09:20:01.5087364Z Instance Type: m7i-flex.8xlarge 2025-12-04T09:20:01.5087564Z AMI Name: unknown 2025-12-04T09:20:01.5126866Z AMI ID: ami-08982f1c5bf93d976 2025-12-04T09:20:05.5833766Z ##[group]Run pytorch/test-infra/.github/actions/setup-ssh@main 2025-12-04T09:20:05.5834068Z with: 2025-12-04T09:20:05.5834608Z github-secret: *** 2025-12-04T09:20:05.5835043Z instructions: All testing is done inside the container, to start an interactive session run: docker exec -it $(docker container ps --format '{{.ID}}') bash 2025-12-04T09:20:05.5835481Z activate-with-label: false 2025-12-04T09:20:05.5835672Z label: with-ssh 2025-12-04T09:20:05.5835845Z remove-existing-keys: true 2025-12-04T09:20:05.5836234Z fail-silently: true 2025-12-04T09:20:05.5836401Z env: 2025-12-04T09:20:05.5836556Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:20:05.5836748Z ##[endgroup] 2025-12-04T09:20:05.7036096Z Please see https://github.com/pytorch/pytorch/wiki/Debugging-using-with-ssh-for-Github-Actions for more info. 2025-12-04T09:20:05.7037986Z Not on pull request and ciflow reference could not be extracted, skipping adding ssh keys 2025-12-04T09:20:05.7161403Z ##[group]Run pytorch/pytorch/.github/actions/checkout-pytorch@main 2025-12-04T09:20:05.7161679Z with: 2025-12-04T09:20:05.7162026Z no-sudo: true 2025-12-04T09:20:05.7162244Z submodules: recursive 2025-12-04T09:20:05.7162492Z fetch-depth: 0 2025-12-04T09:20:05.7162706Z env: 2025-12-04T09:20:05.7162899Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:20:05.7163137Z ##[endgroup] 2025-12-04T09:20:05.7222053Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:20:05.7222615Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:20:05.7229672Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:20:05.7229910Z env: 2025-12-04T09:20:05.7230084Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:20:05.7230505Z ##[endgroup] 2025-12-04T09:20:05.7302507Z ##[group]Run # Use all available CPUs for fetching 2025-12-04T09:20:05.7302814Z # Use all available CPUs for fetching 2025-12-04T09:20:05.7303030Z cd "${GITHUB_WORKSPACE}" 2025-12-04T09:20:05.7303255Z git config --global fetch.parallel 0 2025-12-04T09:20:05.7303499Z git config --global submodule.fetchJobs 0 2025-12-04T09:20:05.7303705Z  2025-12-04T09:20:05.7304027Z # Clean workspace. The default checkout action should also do this, but 2025-12-04T09:20:05.7304322Z # do it here as well just in case 2025-12-04T09:20:05.7304538Z if [[ -d .git ]]; then 2025-12-04T09:20:05.7304724Z  if [ -z "${NO_SUDO}" ]; then 2025-12-04T09:20:05.7304920Z  sudo git clean -ffdx 2025-12-04T09:20:05.7305100Z  else 2025-12-04T09:20:05.7305252Z  git clean -ffdx 2025-12-04T09:20:05.7305420Z  fi 2025-12-04T09:20:05.7305565Z fi 2025-12-04T09:20:05.7309619Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:20:05.7309857Z env: 2025-12-04T09:20:05.7310010Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:20:05.7310182Z NO_SUDO: true 2025-12-04T09:20:05.7310358Z ##[endgroup] 2025-12-04T09:20:05.7409336Z ##[group]Run actions/checkout@v4 2025-12-04T09:20:05.7409573Z with: 2025-12-04T09:20:05.7409769Z ref: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:20:05.7409994Z fetch-depth: 0 2025-12-04T09:20:05.7410172Z submodules: recursive 2025-12-04T09:20:05.7410360Z show-progress: false 2025-12-04T09:20:05.7410540Z repository: pytorch/pytorch 2025-12-04T09:20:05.7410878Z token: *** 2025-12-04T09:20:05.7411040Z ssh-strict: true 2025-12-04T09:20:05.7411207Z ssh-user: git 2025-12-04T09:20:05.7411372Z persist-credentials: true 2025-12-04T09:20:05.7411565Z clean: true 2025-12-04T09:20:05.7411741Z sparse-checkout-cone-mode: true 2025-12-04T09:20:05.7411939Z fetch-tags: false 2025-12-04T09:20:05.7412108Z lfs: false 2025-12-04T09:20:05.7412275Z set-safe-directory: true 2025-12-04T09:20:05.7412458Z env: 2025-12-04T09:20:05.7412615Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:20:05.7412795Z ##[endgroup] 2025-12-04T09:20:05.8292919Z Syncing repository: pytorch/pytorch 2025-12-04T09:20:05.8293891Z ##[group]Getting Git version info 2025-12-04T09:20:05.8294201Z Working directory is '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-12-04T09:20:05.8294622Z [command]/usr/bin/git version 2025-12-04T09:20:05.8551107Z git version 2.50.1 2025-12-04T09:20:05.8577253Z ##[endgroup] 2025-12-04T09:20:05.8581749Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/662b2b49-d7e3-47d1-8926-a3353a54bf18/.gitconfig' 2025-12-04T09:20:05.8618486Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/662b2b49-d7e3-47d1-8926-a3353a54bf18' before making global git config changes 2025-12-04T09:20:05.8620697Z Adding repository directory to the temporary git global config as a safe directory 2025-12-04T09:20:05.8626405Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:20:05.8694242Z Deleting the contents of '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-12-04T09:20:05.8699053Z ##[group]Initializing the repository 2025-12-04T09:20:05.8704521Z [command]/usr/bin/git init /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:20:05.8778573Z hint: Using 'master' as the name for the initial branch. This default branch name 2025-12-04T09:20:05.8783837Z hint: is subject to change. To configure the initial branch name to use in all 2025-12-04T09:20:05.8788986Z hint: of your new repositories, which will suppress this warning, call: 2025-12-04T09:20:05.8793738Z hint: 2025-12-04T09:20:05.8798296Z hint: git config --global init.defaultBranch 2025-12-04T09:20:05.8800408Z hint: 2025-12-04T09:20:05.8800751Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2025-12-04T09:20:05.8801137Z hint: 'development'. The just-created branch can be renamed via this command: 2025-12-04T09:20:05.8801416Z hint: 2025-12-04T09:20:05.8801592Z hint: git branch -m 2025-12-04T09:20:05.8801902Z hint: 2025-12-04T09:20:05.8802174Z hint: Disable this message with "git config set advice.defaultBranchName false" 2025-12-04T09:20:05.8805008Z Initialized empty Git repository in /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/ 2025-12-04T09:20:05.8816660Z [command]/usr/bin/git remote add origin https://github.com/pytorch/pytorch 2025-12-04T09:20:05.8874173Z ##[endgroup] 2025-12-04T09:20:05.8876946Z ##[group]Disabling automatic garbage collection 2025-12-04T09:20:05.8877435Z [command]/usr/bin/git config --local gc.auto 0 2025-12-04T09:20:05.8908527Z ##[endgroup] 2025-12-04T09:20:05.8913095Z ##[group]Setting up auth 2025-12-04T09:20:05.8918275Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-12-04T09:20:05.8940229Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-12-04T09:20:05.9403642Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-12-04T09:20:05.9426366Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-12-04T09:20:05.9762097Z [command]/usr/bin/git config --local --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:20:05.9795587Z [command]/usr/bin/git submodule foreach --recursive git config --local --show-origin --name-only --get-regexp remote.origin.url 2025-12-04T09:20:06.0117327Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-12-04T09:20:06.0190367Z ##[endgroup] 2025-12-04T09:20:06.0196033Z ##[group]Fetching the repository 2025-12-04T09:20:06.0198606Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2025-12-04T09:20:54.1497694Z From https://github.com/pytorch/pytorch 2025-12-04T09:20:54.1498474Z * [new branch] 2.6.0.dev20241004+ -> origin/2.6.0.dev20241004+ 2025-12-04T09:20:54.1499631Z * [new branch] 2.9.1 -> origin/2.9.1 2025-12-04T09:20:54.1500218Z * [new branch] AaronWang04_addmmfusion_perftest -> origin/AaronWang04_addmmfusion_perftest 2025-12-04T09:20:54.1500681Z * [new branch] Flamefire-patch-1 -> origin/Flamefire-patch-1 2025-12-04T09:20:54.1502244Z * [new branch] HDCharles-2.6.0-release-notes -> origin/HDCharles-2.6.0-release-notes 2025-12-04T09:20:54.1502653Z * [new branch] HOPrintFunc -> origin/HOPrintFunc 2025-12-04T09:20:54.1503044Z * [new branch] IvanKobzarev/stack/1 -> origin/IvanKobzarev/stack/1 2025-12-04T09:20:54.1503550Z * [new branch] NicoshevSVE128 -> origin/NicoshevSVE128 2025-12-04T09:20:54.1503923Z * [new branch] PR-AOTInductorNoneBug -> origin/PR-AOTInductorNoneBug 2025-12-04T09:20:54.1504338Z * [new branch] PR-AOTInductorNoneBugFix -> origin/PR-AOTInductorNoneBugFix 2025-12-04T09:20:54.1504747Z * [new branch] PR-FixConfigsIssue -> origin/PR-FixConfigsIssue 2025-12-04T09:20:54.1505129Z * [new branch] PR-NoneBugFix-viable -> origin/PR-NoneBugFix-viable 2025-12-04T09:20:54.1505502Z * [new branch] PR-ResetToZero -> origin/PR-ResetToZero 2025-12-04T09:20:54.1505882Z * [new branch] Update-Flash-Packaging -> origin/Update-Flash-Packaging 2025-12-04T09:20:54.1506258Z * [new branch] VLA_exp -> origin/VLA_exp 2025-12-04T09:20:54.1506602Z * [new branch] activation_bench -> origin/activation_bench 2025-12-04T09:20:54.1509378Z * [new branch] addmm-heuristic -> origin/addmm-heuristic 2025-12-04T09:20:54.1509804Z * [new branch] adi/onednn_aarch64 -> origin/adi/onednn_aarch64 2025-12-04T09:20:54.1510211Z * [new branch] adi/test -> origin/adi/test 2025-12-04T09:20:54.1510552Z * [new branch] adi/test_bgemm -> origin/adi/test_bgemm 2025-12-04T09:20:54.1510885Z * [new branch] adi/test_m8g -> origin/adi/test_m8g 2025-12-04T09:20:54.1511283Z * [new branch] adi/test_onednn -> origin/adi/test_onednn 2025-12-04T09:20:54.1511666Z * [new branch] adi/test_onednn_v3.9 -> origin/adi/test_onednn_v3.9 2025-12-04T09:20:54.1512626Z * [new branch] adi/test_presve_change -> origin/adi/test_presve_change 2025-12-04T09:20:54.1513034Z * [new branch] adi/test_timm -> origin/adi/test_timm 2025-12-04T09:20:54.1513722Z * [new branch] adi/testpresve_change -> origin/adi/testpresve_change 2025-12-04T09:20:54.1515353Z * [new branch] aditew01/test/vec_bf16 -> origin/aditew01/test/vec_bf16 2025-12-04T09:20:54.1516318Z * [new branch] ah-globalfeedback-hook -> origin/ah-globalfeedback-hook 2025-12-04T09:20:54.1517159Z * [new branch] albanD-patch-1 -> origin/albanD-patch-1 2025-12-04T09:20:54.1517771Z * [new branch] also-surround-shimh -> origin/also-surround-shimh 2025-12-04T09:20:54.1518626Z * [new branch] angelayi/aot_compile -> origin/angelayi/aot_compile 2025-12-04T09:20:54.1519437Z * [new branch] angelayi/aoti_additional_files -> origin/angelayi/aoti_additional_files 2025-12-04T09:20:54.1520171Z * [new branch] angelayi/benchmark -> origin/angelayi/benchmark 2025-12-04T09:20:54.1521059Z * [new branch] angelayi/change_pytree_serialization -> origin/angelayi/change_pytree_serialization 2025-12-04T09:20:54.1521846Z * [new branch] angelayi/cpp_loader -> origin/angelayi/cpp_loader 2025-12-04T09:20:54.1522617Z * [new branch] angelayi/inductor_const -> origin/angelayi/inductor_const 2025-12-04T09:20:54.1523324Z * [new branch] angelayi/lstm -> origin/angelayi/lstm 2025-12-04T09:20:54.1524708Z * [new branch] angelayi/no_so_weight -> origin/angelayi/no_so_weight 2025-12-04T09:20:54.1525349Z * [new branch] angelayi/scan_layers -> origin/angelayi/scan_layers 2025-12-04T09:20:54.1526197Z * [new branch] angelayi/side_eff -> origin/angelayi/side_eff 2025-12-04T09:20:54.1526876Z * [new branch] angelayi/state_dict -> origin/angelayi/state_dict 2025-12-04T09:20:54.1527690Z * [new branch] angelayi/symint_input -> origin/angelayi/symint_input 2025-12-04T09:20:54.1528600Z * [new branch] angelayi/symm_mem -> origin/angelayi/symm_mem 2025-12-04T09:20:54.1529217Z * [new branch] angelayi/test_cpp -> origin/angelayi/test_cpp 2025-12-04T09:20:54.1529963Z * [new branch] angelayi/torch_size -> origin/angelayi/torch_size 2025-12-04T09:20:54.1530781Z * [new branch] annotate_assert -> origin/annotate_assert 2025-12-04T09:20:54.1532044Z * [new branch] annotate_fallback_kernel -> origin/annotate_fallback_kernel 2025-12-04T09:20:54.1533673Z * [new branch] annotation_deepcopy -> origin/annotation_deepcopy 2025-12-04T09:20:54.1534192Z * [new branch] annotation_dynamo -> origin/annotation_dynamo 2025-12-04T09:20:54.1534562Z * [new branch] aot_eager_stack_trace -> origin/aot_eager_stack_trace 2025-12-04T09:20:54.1534914Z * [new branch] aoti-cuda-alloc -> origin/aoti-cuda-alloc 2025-12-04T09:20:54.1536756Z * [new branch] aoti_const_device -> origin/aoti_const_device 2025-12-04T09:20:54.1537222Z * [new branch] aoti_fqn_name_interface -> origin/aoti_fqn_name_interface 2025-12-04T09:20:54.1537627Z * [new branch] aoti_package_weights_binary -> origin/aoti_package_weights_binary 2025-12-04T09:20:54.1538017Z * [new branch] aoti_target_windows -> origin/aoti_target_windows 2025-12-04T09:20:54.1542553Z * [new branch] arsh/feat/inductor_check_profiling -> origin/arsh/feat/inductor_check_profiling 2025-12-04T09:20:54.1543052Z * [new branch] async_tp -> origin/async_tp 2025-12-04T09:20:54.1543459Z * [new branch] atalman-inductor-perf-cu124 -> origin/atalman-inductor-perf-cu124 2025-12-04T09:20:54.1544094Z * [new branch] atalman-inductor-perf-cu124.1 -> origin/atalman-inductor-perf-cu124.1 2025-12-04T09:20:54.1544551Z * [new branch] atalman-patch-2 -> origin/atalman-patch-2 2025-12-04T09:20:54.1544924Z * [new branch] atalman-patch-3 -> origin/atalman-patch-3 2025-12-04T09:20:54.1545267Z * [new branch] atalman-patch-4 -> origin/atalman-patch-4 2025-12-04T09:20:54.1545612Z * [new branch] atalman-patch-5 -> origin/atalman-patch-5 2025-12-04T09:20:54.1545976Z * [new branch] atalman-patch-6 -> origin/atalman-patch-6 2025-12-04T09:20:54.1550794Z * [new branch] atalman-patch-7 -> origin/atalman-patch-7 2025-12-04T09:20:54.1551375Z * [new branch] atalman-patch-8 -> origin/atalman-patch-8 2025-12-04T09:20:54.1551754Z * [new branch] atalman_inductor_2.3.1 -> origin/atalman_inductor_2.3.1 2025-12-04T09:20:54.1552128Z * [new branch] atalman_inductor_2.4.0 -> origin/atalman_inductor_2.4.0 2025-12-04T09:20:54.1552511Z * [new branch] atalman_inductor_2.4.x -> origin/atalman_inductor_2.4.x 2025-12-04T09:20:54.1552946Z * [new branch] attention_benchmarking_clean -> origin/attention_benchmarking_clean 2025-12-04T09:20:54.1553424Z * [new branch] bahuang/dt_fix_scalar_add -> origin/bahuang/dt_fix_scalar_add 2025-12-04T09:20:54.1553971Z * [new branch] bahuang/fix_debug_mode -> origin/bahuang/fix_debug_mode 2025-12-04T09:20:54.1554521Z * [new branch] bahuang/fix_expand -> origin/bahuang/fix_expand 2025-12-04T09:20:54.1555018Z * [new branch] bahuang/test -> origin/bahuang/test 2025-12-04T09:20:54.1555665Z * [new branch] base/1.5 -> origin/base/1.5 2025-12-04T09:20:54.1556050Z * [new branch] batching_sdpa_efficient_attention -> origin/batching_sdpa_efficient_attention 2025-12-04T09:20:54.1559409Z * [new branch] bench_scaled_mm_ops -> origin/bench_scaled_mm_ops 2025-12-04T09:20:54.1559881Z * [new branch] benchmark-updates -> origin/benchmark-updates 2025-12-04T09:20:54.1560259Z * [new branch] benchmarking-script -> origin/benchmarking-script 2025-12-04T09:20:54.1560628Z * [new branch] bertmaher/pinbump26 -> origin/bertmaher/pinbump26 2025-12-04T09:20:54.1560993Z * [new branch] bertrand/cutlass -> origin/bertrand/cutlass 2025-12-04T09:20:54.1561366Z * [new branch] bf/bug-static-input -> origin/bf/bug-static-input 2025-12-04T09:20:54.1564257Z * [new branch] bf/cg-backend -> origin/bf/cg-backend 2025-12-04T09:20:54.1564692Z * [new branch] bf/cg-nccl-test -> origin/bf/cg-nccl-test 2025-12-04T09:20:54.1565067Z * [new branch] bf/cg-remove-check -> origin/bf/cg-remove-check 2025-12-04T09:20:54.1565470Z * [new branch] bf/clean-torchbench-hf -> origin/bf/clean-torchbench-hf 2025-12-04T09:20:54.1565863Z * [new branch] bf/combo-debug-log -> origin/bf/combo-debug-log 2025-12-04T09:20:54.1566291Z * [new branch] bf/cudagraph -> origin/bf/cudagraph 2025-12-04T09:20:54.1566749Z * [new branch] bf/cudagraph-disable-input-mutation -> origin/bf/cudagraph-disable-input-mutation 2025-12-04T09:20:54.1567437Z * [new branch] bf/cudagraph-enable-input-mutation-support-benchmark -> origin/bf/cudagraph-enable-input-mutation-support-benchmark 2025-12-04T09:20:54.1568040Z * [new branch] bf/cudagraph-partition -> origin/bf/cudagraph-partition 2025-12-04T09:20:54.1568459Z * [new branch] bf/donated-buffer-bench -> origin/bf/donated-buffer-bench 2025-12-04T09:20:54.1568862Z * [new branch] bf/dynamo-partition -> origin/bf/dynamo-partition 2025-12-04T09:20:54.1569222Z * [new branch] bf/lite -> origin/bf/lite 2025-12-04T09:20:54.1569580Z * [new branch] bf/pa-non-divisible -> origin/bf/pa-non-divisible 2025-12-04T09:20:54.1570089Z * [new branch] bf/partition-cache-free-symbols -> origin/bf/partition-cache-free-symbols 2025-12-04T09:20:54.1570800Z * [new branch] bf/partition-memory-plan -> origin/bf/partition-memory-plan 2025-12-04T09:20:54.1571559Z * [new branch] bf/partition-move-cpu -> origin/bf/partition-move-cpu 2025-12-04T09:20:54.1572400Z * [new branch] bf/partition-view-fallback -> origin/bf/partition-view-fallback 2025-12-04T09:20:54.1573062Z * [new branch] bf/remove-check-55b0c39d -> origin/bf/remove-check-55b0c39d 2025-12-04T09:20:54.1573814Z * [new branch] bf/timm-nov-26-2025 -> origin/bf/timm-nov-26-2025 2025-12-04T09:20:54.1574547Z * [new branch] bf/transformer-pin-4-57-3 -> origin/bf/transformer-pin-4-57-3 2025-12-04T09:20:54.1579616Z * [new branch] bisect_perf_hf_T5_3acc6eac492 -> origin/bisect_perf_hf_T5_3acc6eac492 2025-12-04T09:20:54.1580170Z * [new branch] bisect_perf_hf_T5_3fcf66f61fb -> origin/bisect_perf_hf_T5_3fcf66f61fb 2025-12-04T09:20:54.1580618Z * [new branch] bisect_perf_hf_T5_4009d154129 -> origin/bisect_perf_hf_T5_4009d154129 2025-12-04T09:20:54.1581038Z * [new branch] bisect_perf_hf_T5_40d0740e73d -> origin/bisect_perf_hf_T5_40d0740e73d 2025-12-04T09:20:54.1581448Z * [new branch] bisect_perf_hf_T5_5268754e -> origin/bisect_perf_hf_T5_5268754e 2025-12-04T09:20:54.1581864Z * [new branch] bisect_perf_hf_T5_7d89a8d385c -> origin/bisect_perf_hf_T5_7d89a8d385c 2025-12-04T09:20:54.1582484Z * [new branch] bisect_perf_hf_T5_b7a25c1ee7c -> origin/bisect_perf_hf_T5_b7a25c1ee7c 2025-12-04T09:20:54.1582912Z * [new branch] bisect_perf_hf_T5_c25b201583f -> origin/bisect_perf_hf_T5_c25b201583f 2025-12-04T09:20:54.1583333Z * [new branch] bisect_perf_hf_T5_c93e57efac0 -> origin/bisect_perf_hf_T5_c93e57efac0 2025-12-04T09:20:54.1583800Z * [new branch] bisect_perf_hf_T5_ca9813ea149 -> origin/bisect_perf_hf_T5_ca9813ea149 2025-12-04T09:20:54.1584201Z * [new branch] bisect_perf_hf_T5_d65f194a -> origin/bisect_perf_hf_T5_d65f194a 2025-12-04T09:20:54.1584595Z * [new branch] bisect_perf_hf_T5_da94ab0b -> origin/bisect_perf_hf_T5_da94ab0b 2025-12-04T09:20:54.1585014Z * [new branch] bisect_perf_hf_T5_da94ab0b_new -> origin/bisect_perf_hf_T5_da94ab0b_new 2025-12-04T09:20:54.1585415Z * [new branch] bisect_perf_hf_T5_db4e8a1d8a8 -> origin/bisect_perf_hf_T5_db4e8a1d8a8 2025-12-04T09:20:54.1585827Z * [new branch] bisect_perf_hf_T5_e0d97e936a2 -> origin/bisect_perf_hf_T5_e0d97e936a2 2025-12-04T09:20:54.1586232Z * [new branch] bisect_perf_hf_T5_f23621ec563 -> origin/bisect_perf_hf_T5_f23621ec563 2025-12-04T09:20:54.1590829Z * [new branch] brister/fx_device_type -> origin/brister/fx_device_type 2025-12-04T09:20:54.1591319Z * [new branch] brister/test_inductor_all_fx -> origin/brister/test_inductor_all_fx 2025-12-04T09:20:54.1591780Z * [new branch] brister/tiled_reduction_no_numel_check -> origin/brister/tiled_reduction_no_numel_check 2025-12-04T09:20:54.1592201Z * [new branch] bwd-backup -> origin/bwd-backup 2025-12-04T09:20:54.1592530Z * [new branch] c57382a49 -> origin/c57382a49 2025-12-04T09:20:54.1592850Z * [new branch] ca_0431d47eaa -> origin/ca_0431d47eaa 2025-12-04T09:20:54.1593497Z * [new branch] ca_fix_0431d47eaa -> origin/ca_fix_0431d47eaa 2025-12-04T09:20:54.1593899Z * [new branch] camyllh/test_setup_hooks_push -> origin/camyllh/test_setup_hooks_push 2025-12-04T09:20:54.1594315Z * [new branch] cccclai-patch-1 -> origin/cccclai-patch-1 2025-12-04T09:20:54.1598653Z * [new branch] cherry-pick-159969-by-pytorch_bot_bot_ -> origin/cherry-pick-159969-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1599548Z * [new branch] cherry-pick-160586-by-pytorch_bot_bot_ -> origin/cherry-pick-160586-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1600079Z * [new branch] cherry-pick-162208-by-pytorch_bot_bot_ -> origin/cherry-pick-162208-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1600586Z * [new branch] cherry-pick-163169-by-pytorch_bot_bot_ -> origin/cherry-pick-163169-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1601089Z * [new branch] cherry-pick-165086-by-pytorch_bot_bot_ -> origin/cherry-pick-165086-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1601590Z * [new branch] cherry-pick-165514-by-pytorch_bot_bot_ -> origin/cherry-pick-165514-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1602472Z * [new branch] cherry-pick-165601-by-pytorch_bot_bot_ -> origin/cherry-pick-165601-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1603126Z * [new branch] cherry-pick-165667-by-pytorch_bot_bot_ -> origin/cherry-pick-165667-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1603627Z * [new branch] cherry-pick-165815-by-pytorch_bot_bot_ -> origin/cherry-pick-165815-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1604120Z * [new branch] cherry-pick-165922-by-pytorch_bot_bot_ -> origin/cherry-pick-165922-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1604621Z * [new branch] cherry-pick-166148-by-pytorch_bot_bot_ -> origin/cherry-pick-166148-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1609957Z * [new branch] cherry-pick-166181-by-pytorch_bot_bot_ -> origin/cherry-pick-166181-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1610744Z * [new branch] cherry-pick-166404-by-pytorch_bot_bot_ -> origin/cherry-pick-166404-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1611264Z * [new branch] cherry-pick-166427-by-pytorch_bot_bot_ -> origin/cherry-pick-166427-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1611859Z * [new branch] cherry-pick-166480-by-pytorch_bot_bot_ -> origin/cherry-pick-166480-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1612368Z * [new branch] cherry-pick-166570-by-pytorch_bot_bot_ -> origin/cherry-pick-166570-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1612868Z * [new branch] cherry-pick-166993-by-pytorch_bot_bot_ -> origin/cherry-pick-166993-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1613375Z * [new branch] cherry-pick-167111-by-pytorch_bot_bot_ -> origin/cherry-pick-167111-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1613873Z * [new branch] cherry-pick-167478-by-pytorch_bot_bot_ -> origin/cherry-pick-167478-by-pytorch_bot_bot_ 2025-12-04T09:20:54.1614314Z * [new branch] cherry_pick_166036_166040 -> origin/cherry_pick_166036_166040 2025-12-04T09:20:54.1614684Z * [new branch] cherry_pick_166457 -> origin/cherry_pick_166457 2025-12-04T09:20:54.1615027Z * [new branch] cherrypick_166338 -> origin/cherrypick_166338 2025-12-04T09:20:54.1615377Z * [new branch] cherrypick_166458 -> origin/cherrypick_166458 2025-12-04T09:20:54.1615722Z * [new branch] cherrypick_166586 -> origin/cherrypick_166586 2025-12-04T09:20:54.1616258Z * [new branch] cherrypick_166956 -> origin/cherrypick_166956 2025-12-04T09:20:54.1616593Z * [new branch] ci_attn -> origin/ci_attn 2025-12-04T09:20:54.1616934Z * [new branch] codex-testing -> origin/codex-testing 2025-12-04T09:20:54.1617433Z * [new branch] codex/add-check_memory_overlap-helper-functions -> origin/codex/add-check_memory_overlap-helper-functions 2025-12-04T09:20:54.1618013Z * [new branch] codex/fix-issue-121219-in-pytorch -> origin/codex/fix-issue-121219-in-pytorch 2025-12-04T09:20:54.1618578Z * [new branch] codex/investigate-segfaults-in-get_tensor_storage_id -> origin/codex/investigate-segfaults-in-get_tensor_storage_id 2025-12-04T09:20:54.1623872Z * [new branch] codex/refactor-lintrunner-config-to-use-uv-run -> origin/codex/refactor-lintrunner-config-to-use-uv-run 2025-12-04T09:20:54.1630102Z * [new branch] compatiblpy39util -> origin/compatiblpy39util 2025-12-04T09:20:54.1630845Z * [new branch] cond_hop_device -> origin/cond_hop_device 2025-12-04T09:20:54.1631191Z * [new branch] context_test -> origin/context_test 2025-12-04T09:20:54.1631642Z * [new branch] copilot/code-style-cleanup-python-pip -> origin/copilot/code-style-cleanup-python-pip 2025-12-04T09:20:54.1632135Z * [new branch] cpio/fix_new_ami_tests -> origin/cpio/fix_new_ami_tests 2025-12-04T09:20:54.1632538Z * [new branch] cpp-docs-dependency-upgrade -> origin/cpp-docs-dependency-upgrade 2025-12-04T09:20:54.1632993Z * [new branch] crpa/typo-in-inductor_comm_lowering -> origin/crpa/typo-in-inductor_comm_lowering 2025-12-04T09:20:54.1633429Z * [new branch] csl/always_produce_xml -> origin/csl/always_produce_xml 2025-12-04T09:20:54.1633814Z * [new branch] csl/build_test_more_procs -> origin/csl/build_test_more_procs 2025-12-04T09:20:54.1634206Z * [new branch] csl/build_test_more_procs2 -> origin/csl/build_test_more_procs2 2025-12-04T09:20:54.1634565Z * [new branch] csl/clean_up -> origin/csl/clean_up 2025-12-04T09:20:54.1634920Z * [new branch] csl/fix_retry_segfault_exit -> origin/csl/fix_retry_segfault_exit 2025-12-04T09:20:54.1635276Z * [new branch] csl/katex -> origin/csl/katex 2025-12-04T09:20:54.1635859Z * [new branch] csl/larger_runner -> origin/csl/larger_runner 2025-12-04T09:20:54.1636201Z * [new branch] csl/lint_testing -> origin/csl/lint_testing 2025-12-04T09:20:54.1636551Z * [new branch] csl/lint_thing -> origin/csl/lint_thing 2025-12-04T09:20:54.1636978Z * [new branch] csl/lintrunner_stuff -> origin/csl/lintrunner_stuff 2025-12-04T09:20:54.1639281Z * [new branch] csl/manually_gen_json -> origin/csl/manually_gen_json 2025-12-04T09:20:54.1639638Z * [new branch] csl/mps_sharding -> origin/csl/mps_sharding 2025-12-04T09:20:54.1640001Z * [new branch] csl/multistage_docker -> origin/csl/multistage_docker 2025-12-04T09:20:54.1640361Z * [new branch] csl/print_timing -> origin/csl/print_timing 2025-12-04T09:20:54.1640720Z * [new branch] csl/remove_experiment -> origin/csl/remove_experiment 2025-12-04T09:20:54.1641099Z * [new branch] csl/remove_maybe_unused_var -> origin/csl/remove_maybe_unused_var 2025-12-04T09:20:54.1641530Z * [new branch] csl/remove_repo_specific_autolabel -> origin/csl/remove_repo_specific_autolabel 2025-12-04T09:20:54.1642137Z * [new branch] csl/remove_run_parallel -> origin/csl/remove_run_parallel 2025-12-04T09:20:54.1642518Z * [new branch] csl/remove_unused_vars -> origin/csl/remove_unused_vars 2025-12-04T09:20:54.1642871Z * [new branch] csl/revert_open -> origin/csl/revert_open 2025-12-04T09:20:54.1643213Z * [new branch] csl/skip_build -> origin/csl/skip_build 2025-12-04T09:20:54.1643591Z * [new branch] csl/smaller_avx_amx_runenrs -> origin/csl/smaller_avx_amx_runenrs 2025-12-04T09:20:54.1643955Z * [new branch] csl/td_job_level -> origin/csl/td_job_level 2025-12-04T09:20:54.1644344Z * [new branch] csl/test_cuda_build_large_runner -> origin/csl/test_cuda_build_large_runner 2025-12-04T09:20:54.1649356Z * [new branch] csl/test_owners_autograd_dispatch_nn -> origin/csl/test_owners_autograd_dispatch_nn 2025-12-04T09:20:54.1649903Z * [new branch] csl/test_owners_higher_confidence -> origin/csl/test_owners_higher_confidence 2025-12-04T09:20:54.1650356Z * [new branch] csl/upload_json_running -> origin/csl/upload_json_running 2025-12-04T09:20:54.1650802Z * [new branch] csl/win_sccache -> origin/csl/win_sccache 2025-12-04T09:20:54.1651147Z * [new branch] csl/xml_stuff -> origin/csl/xml_stuff 2025-12-04T09:20:54.1651482Z * [new branch] cublasrelax2 -> origin/cublasrelax2 2025-12-04T09:20:54.1651818Z * [new branch] cuda_mempool -> origin/cuda_mempool 2025-12-04T09:20:54.1652170Z * [new branch] custom_lowering_dict -> origin/custom_lowering_dict 2025-12-04T09:20:54.1652745Z * [new branch] d4l3k/debug_plane_frtrace -> origin/d4l3k/debug_plane_frtrace 2025-12-04T09:20:54.1653115Z * [new branch] daxia6/2.8o3 -> origin/daxia6/2.8o3 2025-12-04T09:20:54.1653858Z * [new branch] debug-guard -> origin/debug-guard 2025-12-04T09:20:54.1654309Z * [new branch] delete-quant-docs -> origin/delete-quant-docs 2025-12-04T09:20:54.1654970Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 2025-12-04T09:20:54.1655841Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 2025-12-04T09:20:54.1656926Z * [new branch] desertfire/test_cpp_wrapper -> origin/desertfire/test_cpp_wrapper 2025-12-04T09:20:54.1657674Z * [new branch] desertfire/triton-cpu-for-aarch64 -> origin/desertfire/triton-cpu-for-aarch64 2025-12-04T09:20:54.1658150Z * [new branch] dev/dhruva/flex_attn_opt -> origin/dev/dhruva/flex_attn_opt 2025-12-04T09:20:54.1660474Z * [new branch] dev/joona/MPSNDArrayAdd -> origin/dev/joona/MPSNDArrayAdd 2025-12-04T09:20:54.1660993Z * [new branch] dev/joona/Unranked -> origin/dev/joona/Unranked 2025-12-04T09:20:54.1661477Z * [new branch] dev/joona/cat -> origin/dev/joona/cat 2025-12-04T09:20:54.1665357Z * [new branch] dev/joona/embeddingbag -> origin/dev/joona/embeddingbag 2025-12-04T09:20:54.1665894Z * [new branch] dev/joona/fix_sdpa_memtest -> origin/dev/joona/fix_sdpa_memtest 2025-12-04T09:20:54.1666826Z * [new branch] dev/joona/getTensorsString -> origin/dev/joona/getTensorsString 2025-12-04T09:20:54.1667348Z * [new branch] dev/joona/mps_linear_macos14 -> origin/dev/joona/mps_linear_macos14 2025-12-04T09:20:54.1667768Z * [new branch] dev/joona/scalar_clamp -> origin/dev/joona/scalar_clamp 2025-12-04T09:20:54.1669103Z * [new branch] dev/joona/sdpa -> origin/dev/joona/sdpa 2025-12-04T09:20:54.1669483Z * [new branch] dev/joona/sdpa_api -> origin/dev/joona/sdpa_api 2025-12-04T09:20:54.1669850Z * [new branch] dev/joona/type_inf -> origin/dev/joona/type_inf 2025-12-04T09:20:54.1670237Z * [new branch] dev/joona/ulpAssertClose -> origin/dev/joona/ulpAssertClose 2025-12-04T09:20:54.1673912Z * [new branch] dev/joona/upsize3d -> origin/dev/joona/upsize3d 2025-12-04T09:20:54.1674275Z * [new branch] disp_counter -> origin/disp_counter 2025-12-04T09:20:54.1674639Z * [new branch] divyanshk-patch-1 -> origin/divyanshk-patch-1 2025-12-04T09:20:54.1674985Z * [new branch] docs -> origin/docs 2025-12-04T09:20:54.1675300Z * [new branch] documentation -> origin/documentation 2025-12-04T09:20:54.1675660Z * [new branch] eager_model_benchmarks -> origin/eager_model_benchmarks 2025-12-04T09:20:54.1676074Z * [new branch] embg/test_inductor_ci_control -> origin/embg/test_inductor_ci_control 2025-12-04T09:20:54.1679076Z * [new branch] embg/triton_l2_prefetch_128B -> origin/embg/triton_l2_prefetch_128B 2025-12-04T09:20:54.1679502Z * [new branch] embg/triton_l2_prefetch_256B -> origin/embg/triton_l2_prefetch_256B 2025-12-04T09:20:54.1679877Z * [new branch] eqy-patch-1 -> origin/eqy-patch-1 2025-12-04T09:20:54.1680212Z * [new branch] eqy-patch-2 -> origin/eqy-patch-2 2025-12-04T09:20:54.1680525Z * [new branch] eqy-patch-3 -> origin/eqy-patch-3 2025-12-04T09:20:54.1680850Z * [new branch] eqy-patch-4 -> origin/eqy-patch-4 2025-12-04T09:20:54.1681163Z * [new branch] eqy-patch-5 -> origin/eqy-patch-5 2025-12-04T09:20:54.1681485Z * [new branch] eqy-patch-6 -> origin/eqy-patch-6 2025-12-04T09:20:54.1682061Z * [new branch] exclamaforte/amd-ma -> origin/exclamaforte/amd-ma 2025-12-04T09:20:54.1682558Z * [new branch] exclamaforte/combo-kernels-perf-run -> origin/exclamaforte/combo-kernels-perf-run 2025-12-04T09:20:54.1686523Z * [new branch] exclamaforte/do_bench_refactor -> origin/exclamaforte/do_bench_refactor 2025-12-04T09:20:54.1687075Z * [new branch] exclamaforte/enable-mem-dep-fusion -> origin/exclamaforte/enable-mem-dep-fusion 2025-12-04T09:20:54.1687619Z * [new branch] exclamaforte/fix-exhaustive-autotuning -> origin/exclamaforte/fix-exhaustive-autotuning 2025-12-04T09:20:54.1688364Z * [new branch] exclamaforte/fix-trace-parsing-fx-svg -> origin/exclamaforte/fix-trace-parsing-fx-svg 2025-12-04T09:20:54.1688939Z * [new branch] exclamaforte/force-pointwise-cat-perf-run -> origin/exclamaforte/force-pointwise-cat-perf-run 2025-12-04T09:20:54.1689449Z * [new branch] exclamaforte/fusion-data -> origin/exclamaforte/fusion-data 2025-12-04T09:20:54.1689988Z * [new branch] exclamaforte/gemm-benchmark-run -> origin/exclamaforte/gemm-benchmark-run 2025-12-04T09:20:54.1690504Z * [new branch] exclamaforte/gemm-export-model -> origin/exclamaforte/gemm-export-model 2025-12-04T09:20:54.1690944Z * [new branch] exclamaforte/gemm-model -> origin/exclamaforte/gemm-model 2025-12-04T09:20:54.1691452Z * [new branch] exclamaforte/gemm-model-all-data-collection -> origin/exclamaforte/gemm-model-all-data-collection 2025-12-04T09:20:54.1691975Z * [new branch] exclamaforte/gemm-to-amd -> origin/exclamaforte/gemm-to-amd 2025-12-04T09:20:54.1692379Z * [new branch] exclamaforte/just-gemm-model -> origin/exclamaforte/just-gemm-model 2025-12-04T09:20:54.1697310Z * [new branch] exclamaforte/just-gemm-model-no-refactor -> origin/exclamaforte/just-gemm-model-no-refactor 2025-12-04T09:20:54.1697980Z * [new branch] exclamaforte/profile-diff-algo -> origin/exclamaforte/profile-diff-algo 2025-12-04T09:20:54.1698531Z * [new branch] exclamaforte/profiler-visualization -> origin/exclamaforte/profiler-visualization 2025-12-04T09:20:54.1699033Z * [new branch] exclamaforte/test_cpp_wrapper_mode -> origin/exclamaforte/test_cpp_wrapper_mode 2025-12-04T09:20:54.1699581Z * [new branch] exclamaforte/update-autotune-configs -> origin/exclamaforte/update-autotune-configs 2025-12-04T09:20:54.1700129Z * [new branch] exclamaforte/update-autotune-configs-2 -> origin/exclamaforte/update-autotune-configs-2 2025-12-04T09:20:54.1700587Z * [new branch] exec -> origin/exec 2025-12-04T09:20:54.1700929Z * [new branch] experimental-mosaic -> origin/experimental-mosaic 2025-12-04T09:20:54.1701304Z * [new branch] export-D61047529 -> origin/export-D61047529 2025-12-04T09:20:54.1701849Z * [new branch] export-D71412006 -> origin/export-D71412006 2025-12-04T09:20:54.1702476Z * [new branch] export-D73042989 -> origin/export-D73042989 2025-12-04T09:20:54.1702927Z * [new branch] export-D78957093 -> origin/export-D78957093 2025-12-04T09:20:54.1703844Z * [new branch] export-D78996107 -> origin/export-D78996107 2025-12-04T09:20:54.1704271Z * [new branch] export-D80823877 -> origin/export-D80823877 2025-12-04T09:20:54.1704621Z * [new branch] export-D80958642 -> origin/export-D80958642 2025-12-04T09:20:54.1705969Z * [new branch] export-D81054193 -> origin/export-D81054193 2025-12-04T09:20:54.1706334Z * [new branch] export-D81204584 -> origin/export-D81204584 2025-12-04T09:20:54.1706675Z * [new branch] export-D81429090 -> origin/export-D81429090 2025-12-04T09:20:54.1707007Z * [new branch] export-D82250826 -> origin/export-D82250826 2025-12-04T09:20:54.1707362Z * [new branch] export-D82253817 -> origin/export-D82253817 2025-12-04T09:20:54.1707697Z * [new branch] export-D83541846 -> origin/export-D83541846 2025-12-04T09:20:54.1712329Z * [new branch] export-D83627170 -> origin/export-D83627170 2025-12-04T09:20:54.1712788Z * [new branch] export-D83766701 -> origin/export-D83766701 2025-12-04T09:20:54.1713142Z * [new branch] export-D83768878 -> origin/export-D83768878 2025-12-04T09:20:54.1713494Z * [new branch] export-D83769447 -> origin/export-D83769447 2025-12-04T09:20:54.1714018Z * [new branch] export-D84089824 -> origin/export-D84089824 2025-12-04T09:20:54.1714348Z * [new branch] export-D84213020 -> origin/export-D84213020 2025-12-04T09:20:54.1714696Z * [new branch] export-D84373821 -> origin/export-D84373821 2025-12-04T09:20:54.1715088Z * [new branch] export-D84612194 -> origin/export-D84612194 2025-12-04T09:20:54.1715425Z * [new branch] export-D84890985 -> origin/export-D84890985 2025-12-04T09:20:54.1715749Z * [new branch] export-D85122326 -> origin/export-D85122326 2025-12-04T09:20:54.1716080Z * [new branch] export-D86256198 -> origin/export-D86256198 2025-12-04T09:20:54.1719994Z * [new branch] export-D86460608 -> origin/export-D86460608 2025-12-04T09:20:54.1720414Z * [new branch] export-D86474796 -> origin/export-D86474796 2025-12-04T09:20:54.1720788Z * [new branch] export-D86712396 -> origin/export-D86712396 2025-12-04T09:20:54.1721128Z * [new branch] export-D87022129 -> origin/export-D87022129 2025-12-04T09:20:54.1721462Z * [new branch] export-D87838959 -> origin/export-D87838959 2025-12-04T09:20:54.1721994Z * [new branch] export-D88319437 -> origin/export-D88319437 2025-12-04T09:20:54.1722427Z * [new branch] exported-model-train-idempotent -> origin/exported-model-train-idempotent 2025-12-04T09:20:54.1722878Z * [new branch] ezyang-titan-october -> origin/ezyang-titan-october 2025-12-04T09:20:54.1723272Z * [new branch] ezyang-titan-october2 -> origin/ezyang-titan-october2 2025-12-04T09:20:54.1728206Z * [new branch] ezyang-war -> origin/ezyang-war 2025-12-04T09:20:54.1733669Z * [new branch] ezyang/wip-aot-descriptors -> origin/ezyang/wip-aot-descriptors 2025-12-04T09:20:54.1735276Z * [new branch] fa_u8_brgemm -> origin/fa_u8_brgemm 2025-12-04T09:20:54.1736278Z * [new branch] fadeputr/sequence_fbgemm -> origin/fadeputr/sequence_fbgemm 2025-12-04T09:20:54.1736860Z * [new branch] fastmath_baseline -> origin/fastmath_baseline 2025-12-04T09:20:54.1737397Z * [new branch] fbcode/warm -> origin/fbcode/warm 2025-12-04T09:20:54.1737727Z * [new branch] fca -> origin/fca 2025-12-04T09:20:54.1738055Z * [new branch] fca2_ca5984c -> origin/fca2_ca5984c 2025-12-04T09:20:54.1738375Z * [new branch] fca5 -> origin/fca5 2025-12-04T09:20:54.1738730Z * [new branch] feature/justknobs-cpp -> origin/feature/justknobs-cpp 2025-12-04T09:20:54.1739142Z * [new branch] feature/numa-forkserver -> origin/feature/numa-forkserver 2025-12-04T09:20:54.1739521Z * [new branch] ffast_math_baseline -> origin/ffast_math_baseline 2025-12-04T09:20:54.1739863Z * [new branch] ffast_math_target -> origin/ffast_math_target 2025-12-04T09:20:54.1740209Z * [new branch] findhao/base_commit -> origin/findhao/base_commit 2025-12-04T09:20:54.1740562Z * [new branch] findhao/base_commit1 -> origin/findhao/base_commit1 2025-12-04T09:20:54.1740928Z * [new branch] findhao/multistream2 -> origin/findhao/multistream2 2025-12-04T09:20:54.1741293Z * [new branch] findhao/multistream5 -> origin/findhao/multistream5 2025-12-04T09:20:54.1741642Z * [new branch] findhao/multistream6 -> origin/findhao/multistream6 2025-12-04T09:20:54.1742014Z * [new branch] findhao/operatorbench3 -> origin/findhao/operatorbench3 2025-12-04T09:20:54.1742683Z * [new branch] findhao/operatorbench5 -> origin/findhao/operatorbench5 2025-12-04T09:20:54.1743067Z * [new branch] findhao/tritonparse -> origin/findhao/tritonparse 2025-12-04T09:20:54.1743469Z * [new branch] fix-ck-gemm-template-format -> origin/fix-ck-gemm-template-format 2025-12-04T09:20:54.1743932Z * [new branch] fix-config-ignore -> origin/fix-config-ignore 2025-12-04T09:20:54.1744287Z * [new branch] fix-dict-guard -> origin/fix-dict-guard 2025-12-04T09:20:54.1744627Z * [new branch] fix_addmm_issue -> origin/fix_addmm_issue 2025-12-04T09:20:54.1745003Z * [new branch] fix_amd_missing_cluster_dims -> origin/fix_amd_missing_cluster_dims 2025-12-04T09:20:54.1745533Z * [new branch] fix_bench_bwd_pass -> origin/fix_bench_bwd_pass 2025-12-04T09:20:54.1745925Z * [new branch] fix_mem_profiler_config -> origin/fix_mem_profiler_config 2025-12-04T09:20:54.1746425Z * [new branch] fix_nvrtc_discovery -> origin/fix_nvrtc_discovery 2025-12-04T09:20:54.1747213Z * [new branch] fix_op_runner -> origin/fix_op_runner 2025-12-04T09:20:54.1747616Z * [new branch] fix_ubn_159469 -> origin/fix_ubn_159469 2025-12-04T09:20:54.1748011Z * [new branch] fixes-triage -> origin/fixes-triage 2025-12-04T09:20:54.1748658Z * [new branch] fixflashinfer -> origin/fixflashinfer 2025-12-04T09:20:54.1749389Z * [new branch] flash_decoding_cpu -> origin/flash_decoding_cpu 2025-12-04T09:20:54.1750133Z * [new branch] flex-flash -> origin/flex-flash 2025-12-04T09:20:54.1754758Z * [new branch] flex_attention_functorch_grad -> origin/flex_attention_functorch_grad 2025-12-04T09:20:54.1755371Z * [new branch] flex_flash -> origin/flex_flash 2025-12-04T09:20:54.1755922Z * [new branch] fmassa/fix_memeff_sharding_rule -> origin/fmassa/fix_memeff_sharding_rule 2025-12-04T09:20:54.1756407Z * [new branch] fmassa/tests_comm_compute_scheduler -> origin/fmassa/tests_comm_compute_scheduler 2025-12-04T09:20:54.1756839Z * [new branch] forkserver_fix -> origin/forkserver_fix 2025-12-04T09:20:54.1757610Z * [new branch] fsdp2_trace_rules -> origin/fsdp2_trace_rules 2025-12-04T09:20:54.1758408Z * [new branch] fx_cpp -> origin/fx_cpp 2025-12-04T09:20:54.1758794Z * [new branch] fy/fix-win -> origin/fy/fix-win 2025-12-04T09:20:54.1759156Z * [new branch] galv-patch-1 -> origin/galv-patch-1 2025-12-04T09:20:54.1759616Z * [new branch] galv/cudagraphs-conditional-nodes-4 -> origin/galv/cudagraphs-conditional-nodes-4 2025-12-04T09:20:54.1760265Z * [new branch] georgehong/cmakelists-patch -> origin/georgehong/cmakelists-patch 2025-12-04T09:20:54.1762269Z * [new branch] gh/AlnisM/1/base -> origin/gh/AlnisM/1/base 2025-12-04T09:20:54.1762641Z * [new branch] gh/AlnisM/1/head -> origin/gh/AlnisM/1/head 2025-12-04T09:20:54.1767963Z * [new branch] gh/EikanWang/67/base -> origin/gh/EikanWang/67/base 2025-12-04T09:20:54.1768429Z * [new branch] gh/EikanWang/67/head -> origin/gh/EikanWang/67/head 2025-12-04T09:20:54.1768817Z * [new branch] gh/Gasoonjia/1/base -> origin/gh/Gasoonjia/1/base 2025-12-04T09:20:54.1769198Z * [new branch] gh/Gasoonjia/1/head -> origin/gh/Gasoonjia/1/head 2025-12-04T09:20:54.1773543Z * [new branch] gh/H-Huang/131/base -> origin/gh/H-Huang/131/base 2025-12-04T09:20:54.1774008Z * [new branch] gh/H-Huang/131/head -> origin/gh/H-Huang/131/head 2025-12-04T09:20:54.1774363Z * [new branch] gh/H-Huang/131/orig -> origin/gh/H-Huang/131/orig 2025-12-04T09:20:54.1774949Z * [new branch] gh/H-Huang/132/base -> origin/gh/H-Huang/132/base 2025-12-04T09:20:54.1775298Z * [new branch] gh/H-Huang/132/head -> origin/gh/H-Huang/132/head 2025-12-04T09:20:54.1775649Z * [new branch] gh/H-Huang/132/orig -> origin/gh/H-Huang/132/orig 2025-12-04T09:20:54.1776042Z * [new branch] gh/H-Huang/180/base -> origin/gh/H-Huang/180/base 2025-12-04T09:20:54.1776396Z * [new branch] gh/H-Huang/180/head -> origin/gh/H-Huang/180/head 2025-12-04T09:20:54.1776741Z * [new branch] gh/H-Huang/180/orig -> origin/gh/H-Huang/180/orig 2025-12-04T09:20:54.1777093Z * [new branch] gh/H-Huang/182/base -> origin/gh/H-Huang/182/base 2025-12-04T09:20:54.1781773Z * [new branch] gh/H-Huang/182/head -> origin/gh/H-Huang/182/head 2025-12-04T09:20:54.1782228Z * [new branch] gh/H-Huang/182/orig -> origin/gh/H-Huang/182/orig 2025-12-04T09:20:54.1782591Z * [new branch] gh/H-Huang/226/base -> origin/gh/H-Huang/226/base 2025-12-04T09:20:54.1782942Z * [new branch] gh/H-Huang/226/head -> origin/gh/H-Huang/226/head 2025-12-04T09:20:54.1783288Z * [new branch] gh/H-Huang/226/orig -> origin/gh/H-Huang/226/orig 2025-12-04T09:20:54.1783658Z * [new branch] gh/H-Huang/228/base -> origin/gh/H-Huang/228/base 2025-12-04T09:20:54.1784008Z * [new branch] gh/H-Huang/228/head -> origin/gh/H-Huang/228/head 2025-12-04T09:20:54.1784357Z * [new branch] gh/H-Huang/228/orig -> origin/gh/H-Huang/228/orig 2025-12-04T09:20:54.1784736Z * [new branch] gh/IvanKobzarev/150/base -> origin/gh/IvanKobzarev/150/base 2025-12-04T09:20:54.1785140Z * [new branch] gh/IvanKobzarev/150/head -> origin/gh/IvanKobzarev/150/head 2025-12-04T09:20:54.1788945Z * [new branch] gh/IvanKobzarev/150/orig -> origin/gh/IvanKobzarev/150/orig 2025-12-04T09:20:54.1789404Z * [new branch] gh/IvanKobzarev/157/base -> origin/gh/IvanKobzarev/157/base 2025-12-04T09:20:54.1789804Z * [new branch] gh/IvanKobzarev/157/head -> origin/gh/IvanKobzarev/157/head 2025-12-04T09:20:54.1790208Z * [new branch] gh/IvanKobzarev/157/orig -> origin/gh/IvanKobzarev/157/orig 2025-12-04T09:20:54.1790595Z * [new branch] gh/IvanKobzarev/159/base -> origin/gh/IvanKobzarev/159/base 2025-12-04T09:20:54.1790982Z * [new branch] gh/IvanKobzarev/159/head -> origin/gh/IvanKobzarev/159/head 2025-12-04T09:20:54.1791370Z * [new branch] gh/IvanKobzarev/159/orig -> origin/gh/IvanKobzarev/159/orig 2025-12-04T09:20:54.1791754Z * [new branch] gh/IvanKobzarev/162/base -> origin/gh/IvanKobzarev/162/base 2025-12-04T09:20:54.1792135Z * [new branch] gh/IvanKobzarev/162/head -> origin/gh/IvanKobzarev/162/head 2025-12-04T09:20:54.1792529Z * [new branch] gh/IvanKobzarev/162/orig -> origin/gh/IvanKobzarev/162/orig 2025-12-04T09:20:54.1792902Z * [new branch] gh/IvanKobzarev/163/base -> origin/gh/IvanKobzarev/163/base 2025-12-04T09:20:54.1799244Z * [new branch] gh/IvanKobzarev/163/head -> origin/gh/IvanKobzarev/163/head 2025-12-04T09:20:54.1799725Z * [new branch] gh/IvanKobzarev/163/orig -> origin/gh/IvanKobzarev/163/orig 2025-12-04T09:20:54.1800120Z * [new branch] gh/IvanKobzarev/166/base -> origin/gh/IvanKobzarev/166/base 2025-12-04T09:20:54.1800509Z * [new branch] gh/IvanKobzarev/166/head -> origin/gh/IvanKobzarev/166/head 2025-12-04T09:20:54.1800899Z * [new branch] gh/IvanKobzarev/166/orig -> origin/gh/IvanKobzarev/166/orig 2025-12-04T09:20:54.1801283Z * [new branch] gh/IvanKobzarev/167/base -> origin/gh/IvanKobzarev/167/base 2025-12-04T09:20:54.1802102Z * [new branch] gh/IvanKobzarev/167/head -> origin/gh/IvanKobzarev/167/head 2025-12-04T09:20:54.1802491Z * [new branch] gh/IvanKobzarev/167/orig -> origin/gh/IvanKobzarev/167/orig 2025-12-04T09:20:54.1802877Z * [new branch] gh/IvanKobzarev/168/base -> origin/gh/IvanKobzarev/168/base 2025-12-04T09:20:54.1803310Z * [new branch] gh/IvanKobzarev/168/head -> origin/gh/IvanKobzarev/168/head 2025-12-04T09:20:54.1803705Z * [new branch] gh/IvanKobzarev/168/orig -> origin/gh/IvanKobzarev/168/orig 2025-12-04T09:20:54.1808307Z * [new branch] gh/IvanKobzarev/169/base -> origin/gh/IvanKobzarev/169/base 2025-12-04T09:20:54.1810962Z * [new branch] gh/IvanKobzarev/169/head -> origin/gh/IvanKobzarev/169/head 2025-12-04T09:20:54.1811605Z * [new branch] gh/IvanKobzarev/169/orig -> origin/gh/IvanKobzarev/169/orig 2025-12-04T09:20:54.1817365Z * [new branch] gh/IvanKobzarev/170/base -> origin/gh/IvanKobzarev/170/base 2025-12-04T09:20:54.1817833Z * [new branch] gh/IvanKobzarev/170/head -> origin/gh/IvanKobzarev/170/head 2025-12-04T09:20:54.1818233Z * [new branch] gh/IvanKobzarev/170/orig -> origin/gh/IvanKobzarev/170/orig 2025-12-04T09:20:54.1818640Z * [new branch] gh/IvanKobzarev/171/base -> origin/gh/IvanKobzarev/171/base 2025-12-04T09:20:54.1819035Z * [new branch] gh/IvanKobzarev/171/head -> origin/gh/IvanKobzarev/171/head 2025-12-04T09:20:54.1819419Z * [new branch] gh/IvanKobzarev/171/orig -> origin/gh/IvanKobzarev/171/orig 2025-12-04T09:20:54.1819808Z * [new branch] gh/IvanKobzarev/172/base -> origin/gh/IvanKobzarev/172/base 2025-12-04T09:20:54.1820214Z * [new branch] gh/IvanKobzarev/172/head -> origin/gh/IvanKobzarev/172/head 2025-12-04T09:20:54.1820600Z * [new branch] gh/IvanKobzarev/172/orig -> origin/gh/IvanKobzarev/172/orig 2025-12-04T09:20:54.1820979Z * [new branch] gh/IvanKobzarev/173/base -> origin/gh/IvanKobzarev/173/base 2025-12-04T09:20:54.1821366Z * [new branch] gh/IvanKobzarev/173/head -> origin/gh/IvanKobzarev/173/head 2025-12-04T09:20:54.1821747Z * [new branch] gh/IvanKobzarev/173/orig -> origin/gh/IvanKobzarev/173/orig 2025-12-04T09:20:54.1822134Z * [new branch] gh/IvanKobzarev/174/base -> origin/gh/IvanKobzarev/174/base 2025-12-04T09:20:54.1822511Z * [new branch] gh/IvanKobzarev/174/head -> origin/gh/IvanKobzarev/174/head 2025-12-04T09:20:54.1822902Z * [new branch] gh/IvanKobzarev/174/orig -> origin/gh/IvanKobzarev/174/orig 2025-12-04T09:20:54.1823294Z * [new branch] gh/IvanKobzarev/175/base -> origin/gh/IvanKobzarev/175/base 2025-12-04T09:20:54.1823685Z * [new branch] gh/IvanKobzarev/175/head -> origin/gh/IvanKobzarev/175/head 2025-12-04T09:20:54.1824071Z * [new branch] gh/IvanKobzarev/175/orig -> origin/gh/IvanKobzarev/175/orig 2025-12-04T09:20:54.1824465Z * [new branch] gh/IvanKobzarev/176/base -> origin/gh/IvanKobzarev/176/base 2025-12-04T09:20:54.1824853Z * [new branch] gh/IvanKobzarev/176/head -> origin/gh/IvanKobzarev/176/head 2025-12-04T09:20:54.1825254Z * [new branch] gh/IvanKobzarev/176/orig -> origin/gh/IvanKobzarev/176/orig 2025-12-04T09:20:54.1825629Z * [new branch] gh/IvanKobzarev/177/base -> origin/gh/IvanKobzarev/177/base 2025-12-04T09:20:54.1826011Z * [new branch] gh/IvanKobzarev/177/head -> origin/gh/IvanKobzarev/177/head 2025-12-04T09:20:54.1826401Z * [new branch] gh/IvanKobzarev/177/orig -> origin/gh/IvanKobzarev/177/orig 2025-12-04T09:20:54.1826792Z * [new branch] gh/IvanKobzarev/178/base -> origin/gh/IvanKobzarev/178/base 2025-12-04T09:20:54.1827179Z * [new branch] gh/IvanKobzarev/178/head -> origin/gh/IvanKobzarev/178/head 2025-12-04T09:20:54.1827720Z * [new branch] gh/IvanKobzarev/178/orig -> origin/gh/IvanKobzarev/178/orig 2025-12-04T09:20:54.1828110Z * [new branch] gh/IvanKobzarev/179/base -> origin/gh/IvanKobzarev/179/base 2025-12-04T09:20:54.1828513Z * [new branch] gh/IvanKobzarev/179/head -> origin/gh/IvanKobzarev/179/head 2025-12-04T09:20:54.1828971Z * [new branch] gh/IvanKobzarev/179/orig -> origin/gh/IvanKobzarev/179/orig 2025-12-04T09:20:54.1829707Z * [new branch] gh/IvanKobzarev/180/base -> origin/gh/IvanKobzarev/180/base 2025-12-04T09:20:54.1830584Z * [new branch] gh/IvanKobzarev/180/head -> origin/gh/IvanKobzarev/180/head 2025-12-04T09:20:54.1831416Z * [new branch] gh/IvanKobzarev/180/orig -> origin/gh/IvanKobzarev/180/orig 2025-12-04T09:20:54.1836867Z * [new branch] gh/IvanKobzarev/181/base -> origin/gh/IvanKobzarev/181/base 2025-12-04T09:20:54.1837349Z * [new branch] gh/IvanKobzarev/181/head -> origin/gh/IvanKobzarev/181/head 2025-12-04T09:20:54.1837751Z * [new branch] gh/IvanKobzarev/181/orig -> origin/gh/IvanKobzarev/181/orig 2025-12-04T09:20:54.1838145Z * [new branch] gh/IvanKobzarev/182/base -> origin/gh/IvanKobzarev/182/base 2025-12-04T09:20:54.1838533Z * [new branch] gh/IvanKobzarev/182/head -> origin/gh/IvanKobzarev/182/head 2025-12-04T09:20:54.1838917Z * [new branch] gh/IvanKobzarev/182/orig -> origin/gh/IvanKobzarev/182/orig 2025-12-04T09:20:54.1839310Z * [new branch] gh/IvanKobzarev/183/base -> origin/gh/IvanKobzarev/183/base 2025-12-04T09:20:54.1839693Z * [new branch] gh/IvanKobzarev/183/head -> origin/gh/IvanKobzarev/183/head 2025-12-04T09:20:54.1840106Z * [new branch] gh/IvanKobzarev/183/orig -> origin/gh/IvanKobzarev/183/orig 2025-12-04T09:20:54.1840778Z * [new branch] gh/IvanKobzarev/184/base -> origin/gh/IvanKobzarev/184/base 2025-12-04T09:20:54.1841181Z * [new branch] gh/IvanKobzarev/184/head -> origin/gh/IvanKobzarev/184/head 2025-12-04T09:20:54.1842053Z * [new branch] gh/IvanKobzarev/184/orig -> origin/gh/IvanKobzarev/184/orig 2025-12-04T09:20:54.1845982Z * [new branch] gh/NikhilAPatel/1/base -> origin/gh/NikhilAPatel/1/base 2025-12-04T09:20:54.1846389Z * [new branch] gh/NikhilAPatel/1/head -> origin/gh/NikhilAPatel/1/head 2025-12-04T09:20:54.1846768Z * [new branch] gh/NikhilAPatel/2/base -> origin/gh/NikhilAPatel/2/base 2025-12-04T09:20:54.1847135Z * [new branch] gh/NikhilAPatel/2/head -> origin/gh/NikhilAPatel/2/head 2025-12-04T09:20:54.1849515Z * [new branch] gh/NikhilAPatel/4/base -> origin/gh/NikhilAPatel/4/base 2025-12-04T09:20:54.1850011Z * [new branch] gh/NikhilAPatel/4/head -> origin/gh/NikhilAPatel/4/head 2025-12-04T09:20:54.1855366Z * [new branch] gh/NikhilAPatel/5/base -> origin/gh/NikhilAPatel/5/base 2025-12-04T09:20:54.1855849Z * [new branch] gh/NikhilAPatel/5/head -> origin/gh/NikhilAPatel/5/head 2025-12-04T09:20:54.1856247Z * [new branch] gh/NikhilAPatel/5/orig -> origin/gh/NikhilAPatel/5/orig 2025-12-04T09:20:54.1856676Z * [new branch] gh/PaliC/17/base -> origin/gh/PaliC/17/base 2025-12-04T09:20:54.1857059Z * [new branch] gh/PaliC/17/head -> origin/gh/PaliC/17/head 2025-12-04T09:20:54.1857420Z * [new branch] gh/PaliC/17/orig -> origin/gh/PaliC/17/orig 2025-12-04T09:20:54.1857777Z * [new branch] gh/PaliC/18/base -> origin/gh/PaliC/18/base 2025-12-04T09:20:54.1858124Z * [new branch] gh/PaliC/18/head -> origin/gh/PaliC/18/head 2025-12-04T09:20:54.1858469Z * [new branch] gh/PaliC/18/orig -> origin/gh/PaliC/18/orig 2025-12-04T09:20:54.1859138Z * [new branch] gh/PaliC/20/base -> origin/gh/PaliC/20/base 2025-12-04T09:20:54.1859488Z * [new branch] gh/PaliC/20/head -> origin/gh/PaliC/20/head 2025-12-04T09:20:54.1859841Z * [new branch] gh/PaliC/20/orig -> origin/gh/PaliC/20/orig 2025-12-04T09:20:54.1860297Z * [new branch] gh/PaliC/21/base -> origin/gh/PaliC/21/base 2025-12-04T09:20:54.1860714Z * [new branch] gh/PaliC/21/head -> origin/gh/PaliC/21/head 2025-12-04T09:20:54.1861065Z * [new branch] gh/PaliC/21/orig -> origin/gh/PaliC/21/orig 2025-12-04T09:20:54.1866639Z * [new branch] gh/PaliC/23/base -> origin/gh/PaliC/23/base 2025-12-04T09:20:54.1867067Z * [new branch] gh/PaliC/23/head -> origin/gh/PaliC/23/head 2025-12-04T09:20:54.1867430Z * [new branch] gh/PaliC/23/orig -> origin/gh/PaliC/23/orig 2025-12-04T09:20:54.1867813Z * [new branch] gh/PaliC/24/base -> origin/gh/PaliC/24/base 2025-12-04T09:20:54.1868170Z * [new branch] gh/PaliC/24/head -> origin/gh/PaliC/24/head 2025-12-04T09:20:54.1868523Z * [new branch] gh/PaliC/24/orig -> origin/gh/PaliC/24/orig 2025-12-04T09:20:54.1868968Z * [new branch] gh/PaliC/25/head -> origin/gh/PaliC/25/head 2025-12-04T09:20:54.1869336Z * [new branch] gh/PaliC/25/next -> origin/gh/PaliC/25/next 2025-12-04T09:20:54.1869703Z * [new branch] gh/PaliC/25/orig -> origin/gh/PaliC/25/orig 2025-12-04T09:20:54.1870057Z * [new branch] gh/PaliC/26/head -> origin/gh/PaliC/26/head 2025-12-04T09:20:54.1870400Z * [new branch] gh/PaliC/26/next -> origin/gh/PaliC/26/next 2025-12-04T09:20:54.1870738Z * [new branch] gh/PaliC/26/orig -> origin/gh/PaliC/26/orig 2025-12-04T09:20:54.1871538Z * [new branch] gh/PaliC/27/next -> origin/gh/PaliC/27/next 2025-12-04T09:20:54.1873055Z * [new branch] gh/PaliC/28/head -> origin/gh/PaliC/28/head 2025-12-04T09:20:54.1873409Z * [new branch] gh/PaliC/28/next -> origin/gh/PaliC/28/next 2025-12-04T09:20:54.1873757Z * [new branch] gh/PaliC/28/orig -> origin/gh/PaliC/28/orig 2025-12-04T09:20:54.1874379Z * [new branch] gh/PaliC/29/head -> origin/gh/PaliC/29/head 2025-12-04T09:20:54.1875077Z * [new branch] gh/PaliC/29/next -> origin/gh/PaliC/29/next 2025-12-04T09:20:54.1878993Z * [new branch] gh/PaliC/29/orig -> origin/gh/PaliC/29/orig 2025-12-04T09:20:54.1879503Z * [new branch] gh/PaliC/30/head -> origin/gh/PaliC/30/head 2025-12-04T09:20:54.1879867Z * [new branch] gh/PaliC/30/next -> origin/gh/PaliC/30/next 2025-12-04T09:20:54.1880224Z * [new branch] gh/PaliC/30/orig -> origin/gh/PaliC/30/orig 2025-12-04T09:20:54.1880598Z * [new branch] gh/PaliC/31/head -> origin/gh/PaliC/31/head 2025-12-04T09:20:54.1880965Z * [new branch] gh/PaliC/31/next -> origin/gh/PaliC/31/next 2025-12-04T09:20:54.1884858Z * [new branch] gh/PaliC/31/orig -> origin/gh/PaliC/31/orig 2025-12-04T09:20:54.1889264Z * [new branch] gh/PaulZhang12/25/base -> origin/gh/PaulZhang12/25/base 2025-12-04T09:20:54.1893520Z * [new branch] gh/PaulZhang12/25/head -> origin/gh/PaulZhang12/25/head 2025-12-04T09:20:54.1895893Z * [new branch] gh/PaulZhang12/25/orig -> origin/gh/PaulZhang12/25/orig 2025-12-04T09:20:54.1896427Z * [new branch] gh/PaulZhang12/28/base -> origin/gh/PaulZhang12/28/base 2025-12-04T09:20:54.1901468Z * [new branch] gh/PaulZhang12/28/head -> origin/gh/PaulZhang12/28/head 2025-12-04T09:20:54.1906043Z * [new branch] gh/PaulZhang12/28/orig -> origin/gh/PaulZhang12/28/orig 2025-12-04T09:20:54.1910554Z * [new branch] gh/PaulZhang12/31/base -> origin/gh/PaulZhang12/31/base 2025-12-04T09:20:54.1912578Z * [new branch] gh/PaulZhang12/31/head -> origin/gh/PaulZhang12/31/head 2025-12-04T09:20:54.1913389Z * [new branch] gh/PaulZhang12/31/orig -> origin/gh/PaulZhang12/31/orig 2025-12-04T09:20:54.1913791Z * [new branch] gh/PaulZhang12/37/base -> origin/gh/PaulZhang12/37/base 2025-12-04T09:20:54.1914189Z * [new branch] gh/PaulZhang12/37/head -> origin/gh/PaulZhang12/37/head 2025-12-04T09:20:54.1914580Z * [new branch] gh/PaulZhang12/37/orig -> origin/gh/PaulZhang12/37/orig 2025-12-04T09:20:54.1914969Z * [new branch] gh/PaulZhang12/40/base -> origin/gh/PaulZhang12/40/base 2025-12-04T09:20:54.1915346Z * [new branch] gh/PaulZhang12/40/head -> origin/gh/PaulZhang12/40/head 2025-12-04T09:20:54.1915735Z * [new branch] gh/PaulZhang12/40/orig -> origin/gh/PaulZhang12/40/orig 2025-12-04T09:20:54.1916112Z * [new branch] gh/PaulZhang12/42/base -> origin/gh/PaulZhang12/42/base 2025-12-04T09:20:54.1916489Z * [new branch] gh/PaulZhang12/42/head -> origin/gh/PaulZhang12/42/head 2025-12-04T09:20:54.1916876Z * [new branch] gh/PaulZhang12/43/base -> origin/gh/PaulZhang12/43/base 2025-12-04T09:20:54.1917260Z * [new branch] gh/PaulZhang12/43/head -> origin/gh/PaulZhang12/43/head 2025-12-04T09:20:54.1917641Z * [new branch] gh/PaulZhang12/43/orig -> origin/gh/PaulZhang12/43/orig 2025-12-04T09:20:54.1918020Z * [new branch] gh/PaulZhang12/44/base -> origin/gh/PaulZhang12/44/base 2025-12-04T09:20:54.1918408Z * [new branch] gh/PaulZhang12/44/head -> origin/gh/PaulZhang12/44/head 2025-12-04T09:20:54.1918795Z * [new branch] gh/PaulZhang12/45/base -> origin/gh/PaulZhang12/45/base 2025-12-04T09:20:54.1919181Z * [new branch] gh/PaulZhang12/45/head -> origin/gh/PaulZhang12/45/head 2025-12-04T09:20:54.1919554Z * [new branch] gh/PaulZhang12/45/orig -> origin/gh/PaulZhang12/45/orig 2025-12-04T09:20:54.1919936Z * [new branch] gh/PaulZhang12/46/base -> origin/gh/PaulZhang12/46/base 2025-12-04T09:20:54.1920324Z * [new branch] gh/PaulZhang12/46/head -> origin/gh/PaulZhang12/46/head 2025-12-04T09:20:54.1920712Z * [new branch] gh/PaulZhang12/46/orig -> origin/gh/PaulZhang12/46/orig 2025-12-04T09:20:54.1921094Z * [new branch] gh/PaulZhang12/47/base -> origin/gh/PaulZhang12/47/base 2025-12-04T09:20:54.1921476Z * [new branch] gh/PaulZhang12/47/head -> origin/gh/PaulZhang12/47/head 2025-12-04T09:20:54.1922045Z * [new branch] gh/PaulZhang12/47/orig -> origin/gh/PaulZhang12/47/orig 2025-12-04T09:20:54.1922430Z * [new branch] gh/PaulZhang12/48/base -> origin/gh/PaulZhang12/48/base 2025-12-04T09:20:54.1922810Z * [new branch] gh/PaulZhang12/48/head -> origin/gh/PaulZhang12/48/head 2025-12-04T09:20:54.1923189Z * [new branch] gh/PaulZhang12/48/orig -> origin/gh/PaulZhang12/48/orig 2025-12-04T09:20:54.1923576Z * [new branch] gh/SamGinzburg/11/base -> origin/gh/SamGinzburg/11/base 2025-12-04T09:20:54.1923939Z * [new branch] gh/SamGinzburg/11/head -> origin/gh/SamGinzburg/11/head 2025-12-04T09:20:54.1924345Z * [new branch] gh/SherlockNoMad/1/base -> origin/gh/SherlockNoMad/1/base 2025-12-04T09:20:54.1924754Z * [new branch] gh/SherlockNoMad/1/head -> origin/gh/SherlockNoMad/1/head 2025-12-04T09:20:54.1925156Z * [new branch] gh/SherlockNoMad/10/base -> origin/gh/SherlockNoMad/10/base 2025-12-04T09:20:54.1925538Z * [new branch] gh/SherlockNoMad/10/head -> origin/gh/SherlockNoMad/10/head 2025-12-04T09:20:54.1925984Z * [new branch] gh/SherlockNoMad/10/orig -> origin/gh/SherlockNoMad/10/orig 2025-12-04T09:20:54.1926374Z * [new branch] gh/SherlockNoMad/11/base -> origin/gh/SherlockNoMad/11/base 2025-12-04T09:20:54.1926766Z * [new branch] gh/SherlockNoMad/11/head -> origin/gh/SherlockNoMad/11/head 2025-12-04T09:20:54.1927219Z * [new branch] gh/SherlockNoMad/11/orig -> origin/gh/SherlockNoMad/11/orig 2025-12-04T09:20:54.1927603Z * [new branch] gh/SherlockNoMad/12/base -> origin/gh/SherlockNoMad/12/base 2025-12-04T09:20:54.1927992Z * [new branch] gh/SherlockNoMad/12/head -> origin/gh/SherlockNoMad/12/head 2025-12-04T09:20:54.1928380Z * [new branch] gh/SherlockNoMad/12/orig -> origin/gh/SherlockNoMad/12/orig 2025-12-04T09:20:54.1928761Z * [new branch] gh/SherlockNoMad/15/base -> origin/gh/SherlockNoMad/15/base 2025-12-04T09:20:54.1929143Z * [new branch] gh/SherlockNoMad/15/head -> origin/gh/SherlockNoMad/15/head 2025-12-04T09:20:54.1929533Z * [new branch] gh/SherlockNoMad/15/orig -> origin/gh/SherlockNoMad/15/orig 2025-12-04T09:20:54.1929930Z * [new branch] gh/SherlockNoMad/17/base -> origin/gh/SherlockNoMad/17/base 2025-12-04T09:20:54.1930557Z * [new branch] gh/SherlockNoMad/17/head -> origin/gh/SherlockNoMad/17/head 2025-12-04T09:20:54.1930946Z * [new branch] gh/SherlockNoMad/17/orig -> origin/gh/SherlockNoMad/17/orig 2025-12-04T09:20:54.1931344Z * [new branch] gh/SherlockNoMad/18/base -> origin/gh/SherlockNoMad/18/base 2025-12-04T09:20:54.1931729Z * [new branch] gh/SherlockNoMad/18/head -> origin/gh/SherlockNoMad/18/head 2025-12-04T09:20:54.1932129Z * [new branch] gh/SherlockNoMad/18/orig -> origin/gh/SherlockNoMad/18/orig 2025-12-04T09:20:54.1936279Z * [new branch] gh/SherlockNoMad/19/base -> origin/gh/SherlockNoMad/19/base 2025-12-04T09:20:54.1940759Z * [new branch] gh/SherlockNoMad/19/head -> origin/gh/SherlockNoMad/19/head 2025-12-04T09:20:54.1943027Z * [new branch] gh/SherlockNoMad/19/orig -> origin/gh/SherlockNoMad/19/orig 2025-12-04T09:20:54.1943613Z * [new branch] gh/SherlockNoMad/2/base -> origin/gh/SherlockNoMad/2/base 2025-12-04T09:20:54.1949011Z * [new branch] gh/SherlockNoMad/2/head -> origin/gh/SherlockNoMad/2/head 2025-12-04T09:20:54.1953909Z * [new branch] gh/SherlockNoMad/20/base -> origin/gh/SherlockNoMad/20/base 2025-12-04T09:20:54.1958346Z * [new branch] gh/SherlockNoMad/20/head -> origin/gh/SherlockNoMad/20/head 2025-12-04T09:20:54.1959395Z * [new branch] gh/SherlockNoMad/20/orig -> origin/gh/SherlockNoMad/20/orig 2025-12-04T09:20:54.1959883Z * [new branch] gh/SherlockNoMad/21/base -> origin/gh/SherlockNoMad/21/base 2025-12-04T09:20:54.1960318Z * [new branch] gh/SherlockNoMad/21/head -> origin/gh/SherlockNoMad/21/head 2025-12-04T09:20:54.1960724Z * [new branch] gh/SherlockNoMad/21/orig -> origin/gh/SherlockNoMad/21/orig 2025-12-04T09:20:54.1961150Z * [new branch] gh/SherlockNoMad/3/base -> origin/gh/SherlockNoMad/3/base 2025-12-04T09:20:54.1961569Z * [new branch] gh/SherlockNoMad/3/head -> origin/gh/SherlockNoMad/3/head 2025-12-04T09:20:54.1962264Z * [new branch] gh/SherlockNoMad/4/base -> origin/gh/SherlockNoMad/4/base 2025-12-04T09:20:54.1962681Z * [new branch] gh/SherlockNoMad/4/head -> origin/gh/SherlockNoMad/4/head 2025-12-04T09:20:54.1963084Z * [new branch] gh/SherlockNoMad/5/base -> origin/gh/SherlockNoMad/5/base 2025-12-04T09:20:54.1963491Z * [new branch] gh/SherlockNoMad/5/head -> origin/gh/SherlockNoMad/5/head 2025-12-04T09:20:54.1963908Z * [new branch] gh/Sidharth123-cpu/24/base -> origin/gh/Sidharth123-cpu/24/base 2025-12-04T09:20:54.1964547Z * [new branch] gh/Sidharth123-cpu/25/base -> origin/gh/Sidharth123-cpu/25/base 2025-12-04T09:20:54.1964953Z * [new branch] gh/Sidharth123-cpu/26/base -> origin/gh/Sidharth123-cpu/26/base 2025-12-04T09:20:54.1965366Z * [new branch] gh/Sidharth123-cpu/27/base -> origin/gh/Sidharth123-cpu/27/base 2025-12-04T09:20:54.1965835Z * [new branch] gh/StrongerXi/1/base -> origin/gh/StrongerXi/1/base 2025-12-04T09:20:54.1966211Z * [new branch] gh/StrongerXi/1/head -> origin/gh/StrongerXi/1/head 2025-12-04T09:20:54.1966574Z * [new branch] gh/StrongerXi/71/base -> origin/gh/StrongerXi/71/base 2025-12-04T09:20:54.1966945Z * [new branch] gh/StrongerXi/71/head -> origin/gh/StrongerXi/71/head 2025-12-04T09:20:54.1967328Z * [new branch] gh/StrongerXi/72/base -> origin/gh/StrongerXi/72/base 2025-12-04T09:20:54.1967700Z * [new branch] gh/StrongerXi/72/head -> origin/gh/StrongerXi/72/head 2025-12-04T09:20:54.1968067Z * [new branch] gh/StrongerXi/73/base -> origin/gh/StrongerXi/73/base 2025-12-04T09:20:54.1968442Z * [new branch] gh/StrongerXi/73/head -> origin/gh/StrongerXi/73/head 2025-12-04T09:20:54.1968814Z * [new branch] gh/StrongerXi/73/orig -> origin/gh/StrongerXi/73/orig 2025-12-04T09:20:54.1969182Z * [new branch] gh/XilunWu/160/base -> origin/gh/XilunWu/160/base 2025-12-04T09:20:54.1969551Z * [new branch] gh/XilunWu/160/head -> origin/gh/XilunWu/160/head 2025-12-04T09:20:54.1969920Z * [new branch] gh/XilunWu/160/orig -> origin/gh/XilunWu/160/orig 2025-12-04T09:20:54.1970292Z * [new branch] gh/XilunWu/163/base -> origin/gh/XilunWu/163/base 2025-12-04T09:20:54.1970655Z * [new branch] gh/XilunWu/163/head -> origin/gh/XilunWu/163/head 2025-12-04T09:20:54.1971017Z * [new branch] gh/XilunWu/163/orig -> origin/gh/XilunWu/163/orig 2025-12-04T09:20:54.1971382Z * [new branch] gh/XilunWu/168/base -> origin/gh/XilunWu/168/base 2025-12-04T09:20:54.1971747Z * [new branch] gh/XilunWu/168/head -> origin/gh/XilunWu/168/head 2025-12-04T09:20:54.1972109Z * [new branch] gh/XilunWu/168/orig -> origin/gh/XilunWu/168/orig 2025-12-04T09:20:54.1972602Z * [new branch] gh/XilunWu/169/base -> origin/gh/XilunWu/169/base 2025-12-04T09:20:54.1972977Z * [new branch] gh/XilunWu/169/head -> origin/gh/XilunWu/169/head 2025-12-04T09:20:54.1973336Z * [new branch] gh/XilunWu/169/orig -> origin/gh/XilunWu/169/orig 2025-12-04T09:20:54.1973703Z * [new branch] gh/XilunWu/170/base -> origin/gh/XilunWu/170/base 2025-12-04T09:20:54.1974076Z * [new branch] gh/XilunWu/170/head -> origin/gh/XilunWu/170/head 2025-12-04T09:20:54.1974433Z * [new branch] gh/XilunWu/170/orig -> origin/gh/XilunWu/170/orig 2025-12-04T09:20:54.1974795Z * [new branch] gh/XilunWu/171/base -> origin/gh/XilunWu/171/base 2025-12-04T09:20:54.1975147Z * [new branch] gh/XilunWu/171/head -> origin/gh/XilunWu/171/head 2025-12-04T09:20:54.1975513Z * [new branch] gh/XilunWu/171/orig -> origin/gh/XilunWu/171/orig 2025-12-04T09:20:54.1975862Z * [new branch] gh/XilunWu/173/base -> origin/gh/XilunWu/173/base 2025-12-04T09:20:54.1976222Z * [new branch] gh/XilunWu/173/head -> origin/gh/XilunWu/173/head 2025-12-04T09:20:54.1976574Z * [new branch] gh/XilunWu/173/orig -> origin/gh/XilunWu/173/orig 2025-12-04T09:20:54.1976927Z * [new branch] gh/XilunWu/175/base -> origin/gh/XilunWu/175/base 2025-12-04T09:20:54.1977289Z * [new branch] gh/XilunWu/175/head -> origin/gh/XilunWu/175/head 2025-12-04T09:20:54.1977703Z * [new branch] gh/XilunWu/175/orig -> origin/gh/XilunWu/175/orig 2025-12-04T09:20:54.1980226Z * [new branch] gh/XilunWu/176/base -> origin/gh/XilunWu/176/base 2025-12-04T09:20:54.1980597Z * [new branch] gh/XilunWu/176/head -> origin/gh/XilunWu/176/head 2025-12-04T09:20:54.1981007Z * [new branch] gh/XilunWu/176/orig -> origin/gh/XilunWu/176/orig 2025-12-04T09:20:54.1981382Z * [new branch] gh/XuehaiPan/14/base -> origin/gh/XuehaiPan/14/base 2025-12-04T09:20:54.1981753Z * [new branch] gh/XuehaiPan/14/head -> origin/gh/XuehaiPan/14/head 2025-12-04T09:20:54.1982127Z * [new branch] gh/XuehaiPan/14/orig -> origin/gh/XuehaiPan/14/orig 2025-12-04T09:20:54.1984198Z * [new branch] gh/XuehaiPan/179/base -> origin/gh/XuehaiPan/179/base 2025-12-04T09:20:54.1984586Z * [new branch] gh/XuehaiPan/179/head -> origin/gh/XuehaiPan/179/head 2025-12-04T09:20:54.1984956Z * [new branch] gh/XuehaiPan/179/orig -> origin/gh/XuehaiPan/179/orig 2025-12-04T09:20:54.1985330Z * [new branch] gh/XuehaiPan/249/base -> origin/gh/XuehaiPan/249/base 2025-12-04T09:20:54.1985701Z * [new branch] gh/XuehaiPan/249/head -> origin/gh/XuehaiPan/249/head 2025-12-04T09:20:54.1986071Z * [new branch] gh/XuehaiPan/249/orig -> origin/gh/XuehaiPan/249/orig 2025-12-04T09:20:54.1987555Z * [new branch] gh/XuehaiPan/253/base -> origin/gh/XuehaiPan/253/base 2025-12-04T09:20:54.1987924Z * [new branch] gh/XuehaiPan/253/head -> origin/gh/XuehaiPan/253/head 2025-12-04T09:20:54.1988292Z * [new branch] gh/XuehaiPan/253/orig -> origin/gh/XuehaiPan/253/orig 2025-12-04T09:20:54.1988663Z * [new branch] gh/XuehaiPan/254/base -> origin/gh/XuehaiPan/254/base 2025-12-04T09:20:54.1990654Z * [new branch] gh/XuehaiPan/254/head -> origin/gh/XuehaiPan/254/head 2025-12-04T09:20:54.1991579Z * [new branch] gh/XuehaiPan/254/orig -> origin/gh/XuehaiPan/254/orig 2025-12-04T09:20:54.1992368Z * [new branch] gh/XuehaiPan/255/base -> origin/gh/XuehaiPan/255/base 2025-12-04T09:20:54.1992836Z * [new branch] gh/XuehaiPan/255/head -> origin/gh/XuehaiPan/255/head 2025-12-04T09:20:54.1993218Z * [new branch] gh/XuehaiPan/255/orig -> origin/gh/XuehaiPan/255/orig 2025-12-04T09:20:54.1995924Z * [new branch] gh/XuehaiPan/271/base -> origin/gh/XuehaiPan/271/base 2025-12-04T09:20:54.1996449Z * [new branch] gh/XuehaiPan/271/head -> origin/gh/XuehaiPan/271/head 2025-12-04T09:20:54.1996838Z * [new branch] gh/XuehaiPan/271/orig -> origin/gh/XuehaiPan/271/orig 2025-12-04T09:20:54.1997232Z * [new branch] gh/XuehaiPan/343/base -> origin/gh/XuehaiPan/343/base 2025-12-04T09:20:54.1997620Z * [new branch] gh/XuehaiPan/343/head -> origin/gh/XuehaiPan/343/head 2025-12-04T09:20:54.1997990Z * [new branch] gh/XuehaiPan/343/orig -> origin/gh/XuehaiPan/343/orig 2025-12-04T09:20:54.1998368Z * [new branch] gh/XuehaiPan/347/base -> origin/gh/XuehaiPan/347/base 2025-12-04T09:20:54.1998790Z * [new branch] gh/XuehaiPan/347/head -> origin/gh/XuehaiPan/347/head 2025-12-04T09:20:54.1999591Z * [new branch] gh/XuehaiPan/347/orig -> origin/gh/XuehaiPan/347/orig 2025-12-04T09:20:54.2002560Z * [new branch] gh/XuehaiPan/348/base -> origin/gh/XuehaiPan/348/base 2025-12-04T09:20:54.2002948Z * [new branch] gh/XuehaiPan/348/head -> origin/gh/XuehaiPan/348/head 2025-12-04T09:20:54.2003353Z * [new branch] gh/XuehaiPan/348/orig -> origin/gh/XuehaiPan/348/orig 2025-12-04T09:20:54.2003884Z * [new branch] gh/XuehaiPan/350/base -> origin/gh/XuehaiPan/350/base 2025-12-04T09:20:54.2007572Z * [new branch] gh/XuehaiPan/350/head -> origin/gh/XuehaiPan/350/head 2025-12-04T09:20:54.2008003Z * [new branch] gh/XuehaiPan/350/orig -> origin/gh/XuehaiPan/350/orig 2025-12-04T09:20:54.2008570Z * [new branch] gh/XuehaiPan/365/base -> origin/gh/XuehaiPan/365/base 2025-12-04T09:20:54.2008957Z * [new branch] gh/XuehaiPan/365/head -> origin/gh/XuehaiPan/365/head 2025-12-04T09:20:54.2009325Z * [new branch] gh/XuehaiPan/365/orig -> origin/gh/XuehaiPan/365/orig 2025-12-04T09:20:54.2009708Z * [new branch] gh/XuehaiPan/366/base -> origin/gh/XuehaiPan/366/base 2025-12-04T09:20:54.2010085Z * [new branch] gh/XuehaiPan/366/head -> origin/gh/XuehaiPan/366/head 2025-12-04T09:20:54.2013906Z * [new branch] gh/XuehaiPan/370/base -> origin/gh/XuehaiPan/370/base 2025-12-04T09:20:54.2014284Z * [new branch] gh/XuehaiPan/370/head -> origin/gh/XuehaiPan/370/head 2025-12-04T09:20:54.2014679Z * [new branch] gh/XuehaiPan/370/orig -> origin/gh/XuehaiPan/370/orig 2025-12-04T09:20:54.2015059Z * [new branch] gh/XuehaiPan/390/base -> origin/gh/XuehaiPan/390/base 2025-12-04T09:20:54.2015426Z * [new branch] gh/XuehaiPan/390/head -> origin/gh/XuehaiPan/390/head 2025-12-04T09:20:54.2015787Z * [new branch] gh/XuehaiPan/390/orig -> origin/gh/XuehaiPan/390/orig 2025-12-04T09:20:54.2018936Z * [new branch] gh/XuehaiPan/391/base -> origin/gh/XuehaiPan/391/base 2025-12-04T09:20:54.2019444Z * [new branch] gh/XuehaiPan/391/head -> origin/gh/XuehaiPan/391/head 2025-12-04T09:20:54.2019846Z * [new branch] gh/XuehaiPan/391/orig -> origin/gh/XuehaiPan/391/orig 2025-12-04T09:20:54.2020229Z * [new branch] gh/XuehaiPan/392/base -> origin/gh/XuehaiPan/392/base 2025-12-04T09:20:54.2020645Z * [new branch] gh/XuehaiPan/392/head -> origin/gh/XuehaiPan/392/head 2025-12-04T09:20:54.2021030Z * [new branch] gh/XuehaiPan/392/orig -> origin/gh/XuehaiPan/392/orig 2025-12-04T09:20:54.2022330Z * [new branch] gh/XuehaiPan/394/base -> origin/gh/XuehaiPan/394/base 2025-12-04T09:20:54.2022766Z * [new branch] gh/XuehaiPan/394/head -> origin/gh/XuehaiPan/394/head 2025-12-04T09:20:54.2023180Z * [new branch] gh/XuehaiPan/394/orig -> origin/gh/XuehaiPan/394/orig 2025-12-04T09:20:54.2023562Z * [new branch] gh/XuehaiPan/397/base -> origin/gh/XuehaiPan/397/base 2025-12-04T09:20:54.2023980Z * [new branch] gh/XuehaiPan/397/head -> origin/gh/XuehaiPan/397/head 2025-12-04T09:20:54.2028021Z * [new branch] gh/XuehaiPan/397/orig -> origin/gh/XuehaiPan/397/orig 2025-12-04T09:20:54.2028422Z * [new branch] gh/XuehaiPan/398/base -> origin/gh/XuehaiPan/398/base 2025-12-04T09:20:54.2028775Z * [new branch] gh/XuehaiPan/398/head -> origin/gh/XuehaiPan/398/head 2025-12-04T09:20:54.2029133Z * [new branch] gh/XuehaiPan/398/orig -> origin/gh/XuehaiPan/398/orig 2025-12-04T09:20:54.2029485Z * [new branch] gh/XuehaiPan/399/base -> origin/gh/XuehaiPan/399/base 2025-12-04T09:20:54.2029842Z * [new branch] gh/XuehaiPan/399/head -> origin/gh/XuehaiPan/399/head 2025-12-04T09:20:54.2034424Z * [new branch] gh/XuehaiPan/399/orig -> origin/gh/XuehaiPan/399/orig 2025-12-04T09:20:54.2038674Z * [new branch] gh/XuehaiPan/400/base -> origin/gh/XuehaiPan/400/base 2025-12-04T09:20:54.2039111Z * [new branch] gh/XuehaiPan/400/head -> origin/gh/XuehaiPan/400/head 2025-12-04T09:20:54.2039490Z * [new branch] gh/XuehaiPan/400/orig -> origin/gh/XuehaiPan/400/orig 2025-12-04T09:20:54.2040116Z * [new branch] gh/ZhiweiYan-96/39/base -> origin/gh/ZhiweiYan-96/39/base 2025-12-04T09:20:54.2040514Z * [new branch] gh/ZhiweiYan-96/39/head -> origin/gh/ZhiweiYan-96/39/head 2025-12-04T09:20:54.2040892Z * [new branch] gh/ZhiweiYan-96/39/orig -> origin/gh/ZhiweiYan-96/39/orig 2025-12-04T09:20:54.2041340Z * [new branch] gh/ZhiweiYan-96/44/base -> origin/gh/ZhiweiYan-96/44/base 2025-12-04T09:20:54.2041837Z * [new branch] gh/ZhiweiYan-96/44/head -> origin/gh/ZhiweiYan-96/44/head 2025-12-04T09:20:54.2042284Z * [new branch] gh/ZhiweiYan-96/45/base -> origin/gh/ZhiweiYan-96/45/base 2025-12-04T09:20:54.2042645Z * [new branch] gh/ZhiweiYan-96/45/head -> origin/gh/ZhiweiYan-96/45/head 2025-12-04T09:20:54.2043014Z * [new branch] gh/ZhiweiYan-96/49/base -> origin/gh/ZhiweiYan-96/49/base 2025-12-04T09:20:54.2043380Z * [new branch] gh/ZhiweiYan-96/49/head -> origin/gh/ZhiweiYan-96/49/head 2025-12-04T09:20:54.2043747Z * [new branch] gh/ZhiweiYan-96/62/base -> origin/gh/ZhiweiYan-96/62/base 2025-12-04T09:20:54.2044111Z * [new branch] gh/ZhiweiYan-96/62/head -> origin/gh/ZhiweiYan-96/62/head 2025-12-04T09:20:54.2044468Z * [new branch] gh/ZhiweiYan-96/66/base -> origin/gh/ZhiweiYan-96/66/base 2025-12-04T09:20:54.2044826Z * [new branch] gh/ZhiweiYan-96/66/head -> origin/gh/ZhiweiYan-96/66/head 2025-12-04T09:20:54.2045174Z * [new branch] gh/ZhiweiYan-96/67/base -> origin/gh/ZhiweiYan-96/67/base 2025-12-04T09:20:54.2045528Z * [new branch] gh/ZhiweiYan-96/67/head -> origin/gh/ZhiweiYan-96/67/head 2025-12-04T09:20:54.2047008Z * [new branch] gh/ZhiweiYan-96/68/base -> origin/gh/ZhiweiYan-96/68/base 2025-12-04T09:20:54.2047399Z * [new branch] gh/ZhiweiYan-96/68/head -> origin/gh/ZhiweiYan-96/68/head 2025-12-04T09:20:54.2047760Z * [new branch] gh/ZhiweiYan-96/68/orig -> origin/gh/ZhiweiYan-96/68/orig 2025-12-04T09:20:54.2048135Z * [new branch] gh/aakhundov/1/base -> origin/gh/aakhundov/1/base 2025-12-04T09:20:54.2048497Z * [new branch] gh/aakhundov/1/head -> origin/gh/aakhundov/1/head 2025-12-04T09:20:54.2048851Z * [new branch] gh/aakhundov/2/base -> origin/gh/aakhundov/2/base 2025-12-04T09:20:54.2049190Z * [new branch] gh/aakhundov/2/head -> origin/gh/aakhundov/2/head 2025-12-04T09:20:54.2054280Z * [new branch] gh/aditew01/openblas -> origin/gh/aditew01/openblas 2025-12-04T09:20:54.2059286Z * [new branch] gh/aditew01/sbgemm -> origin/gh/aditew01/sbgemm 2025-12-04T09:20:54.2063101Z * [new branch] gh/aditew01/vecbf16 -> origin/gh/aditew01/vecbf16 2025-12-04T09:20:54.2065726Z * [new branch] gh/albanD/4/base -> origin/gh/albanD/4/base 2025-12-04T09:20:54.2066119Z * [new branch] gh/albanD/4/head -> origin/gh/albanD/4/head 2025-12-04T09:20:54.2066796Z * [new branch] gh/albanD/4/orig -> origin/gh/albanD/4/orig 2025-12-04T09:20:54.2067390Z * [new branch] gh/alexbrauckmann/paddedtensor_faketensor_init -> origin/gh/alexbrauckmann/paddedtensor_faketensor_init 2025-12-04T09:20:54.2067946Z * [new branch] gh/alexsamardzic/12/base -> origin/gh/alexsamardzic/12/base 2025-12-04T09:20:54.2068362Z * [new branch] gh/alexsamardzic/12/head -> origin/gh/alexsamardzic/12/head 2025-12-04T09:20:54.2068760Z * [new branch] gh/alexsamardzic/12/orig -> origin/gh/alexsamardzic/12/orig 2025-12-04T09:20:54.2069157Z * [new branch] gh/alexsamardzic/14/base -> origin/gh/alexsamardzic/14/base 2025-12-04T09:20:54.2069548Z * [new branch] gh/alexsamardzic/14/head -> origin/gh/alexsamardzic/14/head 2025-12-04T09:20:54.2070097Z * [new branch] gh/alexsamardzic/14/orig -> origin/gh/alexsamardzic/14/orig 2025-12-04T09:20:54.2070503Z * [new branch] gh/alexsamardzic/15/base -> origin/gh/alexsamardzic/15/base 2025-12-04T09:20:54.2070906Z * [new branch] gh/alexsamardzic/15/head -> origin/gh/alexsamardzic/15/head 2025-12-04T09:20:54.2071358Z * [new branch] gh/alexsamardzic/15/orig -> origin/gh/alexsamardzic/15/orig 2025-12-04T09:20:54.2071744Z * [new branch] gh/amjames/18/base -> origin/gh/amjames/18/base 2025-12-04T09:20:54.2072119Z * [new branch] gh/amjames/18/head -> origin/gh/amjames/18/head 2025-12-04T09:20:54.2072481Z * [new branch] gh/amjames/18/orig -> origin/gh/amjames/18/orig 2025-12-04T09:20:54.2072892Z * [new branch] gh/andrewor14/35/base -> origin/gh/andrewor14/35/base 2025-12-04T09:20:54.2073276Z * [new branch] gh/andrewor14/35/head -> origin/gh/andrewor14/35/head 2025-12-04T09:20:54.2073660Z * [new branch] gh/andrewor14/35/orig -> origin/gh/andrewor14/35/orig 2025-12-04T09:20:54.2074032Z * [new branch] gh/andrewor14/50/base -> origin/gh/andrewor14/50/base 2025-12-04T09:20:54.2074405Z * [new branch] gh/andrewor14/50/head -> origin/gh/andrewor14/50/head 2025-12-04T09:20:54.2074780Z * [new branch] gh/andrewor14/50/orig -> origin/gh/andrewor14/50/orig 2025-12-04T09:20:54.2075150Z * [new branch] gh/andyanwang/30/base -> origin/gh/andyanwang/30/base 2025-12-04T09:20:54.2075516Z * [new branch] gh/andyanwang/30/orig -> origin/gh/andyanwang/30/orig 2025-12-04T09:20:54.2075904Z * [new branch] gh/andyanwang/31/base -> origin/gh/andyanwang/31/base 2025-12-04T09:20:54.2076906Z * [new branch] gh/andyanwang/31/orig -> origin/gh/andyanwang/31/orig 2025-12-04T09:20:54.2077351Z * [new branch] gh/andyanwang/39/base -> origin/gh/andyanwang/39/base 2025-12-04T09:20:54.2077759Z * [new branch] gh/andyanwang/39/head -> origin/gh/andyanwang/39/head 2025-12-04T09:20:54.2078121Z * [new branch] gh/andyanwang/39/orig -> origin/gh/andyanwang/39/orig 2025-12-04T09:20:54.2084865Z * [new branch] gh/andyanwang/42/base -> origin/gh/andyanwang/42/base 2025-12-04T09:20:54.2085322Z * [new branch] gh/andyanwang/42/head -> origin/gh/andyanwang/42/head 2025-12-04T09:20:54.2085715Z * [new branch] gh/andyanwang/42/orig -> origin/gh/andyanwang/42/orig 2025-12-04T09:20:54.2086081Z * [new branch] gh/andyanwang/45/base -> origin/gh/andyanwang/45/base 2025-12-04T09:20:54.2086452Z * [new branch] gh/andyanwang/45/head -> origin/gh/andyanwang/45/head 2025-12-04T09:20:54.2086821Z * [new branch] gh/andyanwang/45/orig -> origin/gh/andyanwang/45/orig 2025-12-04T09:20:54.2087194Z * [new branch] gh/angelayi/107/base -> origin/gh/angelayi/107/base 2025-12-04T09:20:54.2087570Z * [new branch] gh/angelayi/107/head -> origin/gh/angelayi/107/head 2025-12-04T09:20:54.2087938Z * [new branch] gh/angelayi/114/base -> origin/gh/angelayi/114/base 2025-12-04T09:20:54.2088304Z * [new branch] gh/angelayi/114/head -> origin/gh/angelayi/114/head 2025-12-04T09:20:54.2088659Z * [new branch] gh/angelayi/114/orig -> origin/gh/angelayi/114/orig 2025-12-04T09:20:54.2089017Z * [new branch] gh/angelayi/116/base -> origin/gh/angelayi/116/base 2025-12-04T09:20:54.2089372Z * [new branch] gh/angelayi/116/head -> origin/gh/angelayi/116/head 2025-12-04T09:20:54.2089728Z * [new branch] gh/angelayi/116/orig -> origin/gh/angelayi/116/orig 2025-12-04T09:20:54.2093962Z * [new branch] gh/angelayi/122/base -> origin/gh/angelayi/122/base 2025-12-04T09:20:54.2094568Z * [new branch] gh/angelayi/122/head -> origin/gh/angelayi/122/head 2025-12-04T09:20:54.2094942Z * [new branch] gh/angelayi/122/orig -> origin/gh/angelayi/122/orig 2025-12-04T09:20:54.2095306Z * [new branch] gh/angelayi/124/base -> origin/gh/angelayi/124/base 2025-12-04T09:20:54.2095736Z * [new branch] gh/angelayi/124/head -> origin/gh/angelayi/124/head 2025-12-04T09:20:54.2096103Z * [new branch] gh/angelayi/124/orig -> origin/gh/angelayi/124/orig 2025-12-04T09:20:54.2096641Z * [new branch] gh/angelayi/128/base -> origin/gh/angelayi/128/base 2025-12-04T09:20:54.2097013Z * [new branch] gh/angelayi/128/head -> origin/gh/angelayi/128/head 2025-12-04T09:20:54.2097371Z * [new branch] gh/angelayi/128/orig -> origin/gh/angelayi/128/orig 2025-12-04T09:20:54.2097745Z * [new branch] gh/angelayi/131/base -> origin/gh/angelayi/131/base 2025-12-04T09:20:54.2098102Z * [new branch] gh/angelayi/131/head -> origin/gh/angelayi/131/head 2025-12-04T09:20:54.2098449Z * [new branch] gh/angelayi/131/orig -> origin/gh/angelayi/131/orig 2025-12-04T09:20:54.2101238Z * [new branch] gh/angelayi/132/base -> origin/gh/angelayi/132/base 2025-12-04T09:20:54.2101623Z * [new branch] gh/angelayi/132/head -> origin/gh/angelayi/132/head 2025-12-04T09:20:54.2101967Z * [new branch] gh/angelayi/132/orig -> origin/gh/angelayi/132/orig 2025-12-04T09:20:54.2102322Z * [new branch] gh/angelayi/133/base -> origin/gh/angelayi/133/base 2025-12-04T09:20:54.2102670Z * [new branch] gh/angelayi/133/head -> origin/gh/angelayi/133/head 2025-12-04T09:20:54.2103033Z * [new branch] gh/angelayi/133/orig -> origin/gh/angelayi/133/orig 2025-12-04T09:20:54.2105866Z * [new branch] gh/angelayi/134/base -> origin/gh/angelayi/134/base 2025-12-04T09:20:54.2106358Z * [new branch] gh/angelayi/134/head -> origin/gh/angelayi/134/head 2025-12-04T09:20:54.2106849Z * [new branch] gh/angelayi/134/orig -> origin/gh/angelayi/134/orig 2025-12-04T09:20:54.2107345Z * [new branch] gh/angelayi/135/base -> origin/gh/angelayi/135/base 2025-12-04T09:20:54.2107833Z * [new branch] gh/angelayi/135/head -> origin/gh/angelayi/135/head 2025-12-04T09:20:54.2108324Z * [new branch] gh/angelayi/135/orig -> origin/gh/angelayi/135/orig 2025-12-04T09:20:54.2108797Z * [new branch] gh/angelayi/136/base -> origin/gh/angelayi/136/base 2025-12-04T09:20:54.2109480Z * [new branch] gh/angelayi/136/head -> origin/gh/angelayi/136/head 2025-12-04T09:20:54.2109918Z * [new branch] gh/angelayi/136/orig -> origin/gh/angelayi/136/orig 2025-12-04T09:20:54.2111367Z * [new branch] gh/angelayi/137/base -> origin/gh/angelayi/137/base 2025-12-04T09:20:54.2111742Z * [new branch] gh/angelayi/137/head -> origin/gh/angelayi/137/head 2025-12-04T09:20:54.2112113Z * [new branch] gh/angelayi/137/orig -> origin/gh/angelayi/137/orig 2025-12-04T09:20:54.2116391Z * [new branch] gh/angelayi/138/base -> origin/gh/angelayi/138/base 2025-12-04T09:20:54.2117302Z * [new branch] gh/angelayi/138/head -> origin/gh/angelayi/138/head 2025-12-04T09:20:54.2117731Z * [new branch] gh/angelayi/138/orig -> origin/gh/angelayi/138/orig 2025-12-04T09:20:54.2118102Z * [new branch] gh/angelayi/139/base -> origin/gh/angelayi/139/base 2025-12-04T09:20:54.2118473Z * [new branch] gh/angelayi/139/head -> origin/gh/angelayi/139/head 2025-12-04T09:20:54.2119637Z * [new branch] gh/angelayi/139/orig -> origin/gh/angelayi/139/orig 2025-12-04T09:20:54.2120250Z * [new branch] gh/angelayi/140/base -> origin/gh/angelayi/140/base 2025-12-04T09:20:54.2120618Z * [new branch] gh/angelayi/140/head -> origin/gh/angelayi/140/head 2025-12-04T09:20:54.2120980Z * [new branch] gh/angelayi/140/orig -> origin/gh/angelayi/140/orig 2025-12-04T09:20:54.2122099Z * [new branch] gh/angelayi/141/base -> origin/gh/angelayi/141/base 2025-12-04T09:20:54.2122748Z * [new branch] gh/angelayi/141/head -> origin/gh/angelayi/141/head 2025-12-04T09:20:54.2123308Z * [new branch] gh/angelayi/141/orig -> origin/gh/angelayi/141/orig 2025-12-04T09:20:54.2125564Z * [new branch] gh/angelayi/142/base -> origin/gh/angelayi/142/base 2025-12-04T09:20:54.2125942Z * [new branch] gh/angelayi/142/head -> origin/gh/angelayi/142/head 2025-12-04T09:20:54.2126290Z * [new branch] gh/angelayi/142/orig -> origin/gh/angelayi/142/orig 2025-12-04T09:20:54.2126688Z * [new branch] gh/angelayi/143/base -> origin/gh/angelayi/143/base 2025-12-04T09:20:54.2127048Z * [new branch] gh/angelayi/143/head -> origin/gh/angelayi/143/head 2025-12-04T09:20:54.2130470Z * [new branch] gh/angelayi/143/orig -> origin/gh/angelayi/143/orig 2025-12-04T09:20:54.2130832Z * [new branch] gh/angelayi/144/base -> origin/gh/angelayi/144/base 2025-12-04T09:20:54.2131193Z * [new branch] gh/angelayi/144/head -> origin/gh/angelayi/144/head 2025-12-04T09:20:54.2131553Z * [new branch] gh/angelayi/144/orig -> origin/gh/angelayi/144/orig 2025-12-04T09:20:54.2131937Z * [new branch] gh/anijain2305/753/base -> origin/gh/anijain2305/753/base 2025-12-04T09:20:54.2132317Z * [new branch] gh/anijain2305/753/head -> origin/gh/anijain2305/753/head 2025-12-04T09:20:54.2136320Z * [new branch] gh/anijain2305/753/orig -> origin/gh/anijain2305/753/orig 2025-12-04T09:20:54.2136706Z * [new branch] gh/anijain2305/810/base -> origin/gh/anijain2305/810/base 2025-12-04T09:20:54.2137080Z * [new branch] gh/anijain2305/810/head -> origin/gh/anijain2305/810/head 2025-12-04T09:20:54.2137445Z * [new branch] gh/anijain2305/810/orig -> origin/gh/anijain2305/810/orig 2025-12-04T09:20:54.2137818Z * [new branch] gh/anijain2305/854/base -> origin/gh/anijain2305/854/base 2025-12-04T09:20:54.2138185Z * [new branch] gh/anijain2305/854/head -> origin/gh/anijain2305/854/head 2025-12-04T09:20:54.2138554Z * [new branch] gh/anijain2305/854/orig -> origin/gh/anijain2305/854/orig 2025-12-04T09:20:54.2140610Z * [new branch] gh/anijain2305/864/base -> origin/gh/anijain2305/864/base 2025-12-04T09:20:54.2141427Z * [new branch] gh/anijain2305/864/head -> origin/gh/anijain2305/864/head 2025-12-04T09:20:54.2141891Z * [new branch] gh/anijain2305/864/orig -> origin/gh/anijain2305/864/orig 2025-12-04T09:20:54.2142317Z * [new branch] gh/anijain2305/870/base -> origin/gh/anijain2305/870/base 2025-12-04T09:20:54.2142716Z * [new branch] gh/anijain2305/870/head -> origin/gh/anijain2305/870/head 2025-12-04T09:20:54.2143129Z * [new branch] gh/anijain2305/870/orig -> origin/gh/anijain2305/870/orig 2025-12-04T09:20:54.2145198Z * [new branch] gh/anijain2305/873/base -> origin/gh/anijain2305/873/base 2025-12-04T09:20:54.2145580Z * [new branch] gh/anijain2305/873/head -> origin/gh/anijain2305/873/head 2025-12-04T09:20:54.2145955Z * [new branch] gh/anijain2305/873/orig -> origin/gh/anijain2305/873/orig 2025-12-04T09:20:54.2146329Z * [new branch] gh/anijain2305/894/base -> origin/gh/anijain2305/894/base 2025-12-04T09:20:54.2146693Z * [new branch] gh/anijain2305/894/head -> origin/gh/anijain2305/894/head 2025-12-04T09:20:54.2147290Z * [new branch] gh/anijain2305/894/orig -> origin/gh/anijain2305/894/orig 2025-12-04T09:20:54.2150183Z * [new branch] gh/anijain2305/895/base -> origin/gh/anijain2305/895/base 2025-12-04T09:20:54.2150566Z * [new branch] gh/anijain2305/895/head -> origin/gh/anijain2305/895/head 2025-12-04T09:20:54.2151015Z * [new branch] gh/anijain2305/895/orig -> origin/gh/anijain2305/895/orig 2025-12-04T09:20:54.2151379Z * [new branch] gh/anijain2305/910/base -> origin/gh/anijain2305/910/base 2025-12-04T09:20:54.2151747Z * [new branch] gh/anijain2305/910/head -> origin/gh/anijain2305/910/head 2025-12-04T09:20:54.2152117Z * [new branch] gh/anijain2305/910/orig -> origin/gh/anijain2305/910/orig 2025-12-04T09:20:54.2154806Z * [new branch] gh/anijain2305/919/base -> origin/gh/anijain2305/919/base 2025-12-04T09:20:54.2155185Z * [new branch] gh/anijain2305/919/head -> origin/gh/anijain2305/919/head 2025-12-04T09:20:54.2155558Z * [new branch] gh/anijain2305/919/orig -> origin/gh/anijain2305/919/orig 2025-12-04T09:20:54.2155939Z * [new branch] gh/anijain2305/922/base -> origin/gh/anijain2305/922/base 2025-12-04T09:20:54.2156326Z * [new branch] gh/anijain2305/922/head -> origin/gh/anijain2305/922/head 2025-12-04T09:20:54.2156693Z * [new branch] gh/anijain2305/922/orig -> origin/gh/anijain2305/922/orig 2025-12-04T09:20:54.2158517Z * [new branch] gh/anijain2305/932/base -> origin/gh/anijain2305/932/base 2025-12-04T09:20:54.2158901Z * [new branch] gh/anijain2305/932/head -> origin/gh/anijain2305/932/head 2025-12-04T09:20:54.2159283Z * [new branch] gh/anijain2305/932/orig -> origin/gh/anijain2305/932/orig 2025-12-04T09:20:54.2159652Z * [new branch] gh/anijain2305/940/base -> origin/gh/anijain2305/940/base 2025-12-04T09:20:54.2160030Z * [new branch] gh/anijain2305/940/head -> origin/gh/anijain2305/940/head 2025-12-04T09:20:54.2160416Z * [new branch] gh/anijain2305/940/orig -> origin/gh/anijain2305/940/orig 2025-12-04T09:20:54.2161243Z * [new branch] gh/anijain2305/941/base -> origin/gh/anijain2305/941/base 2025-12-04T09:20:54.2162228Z * [new branch] gh/anijain2305/941/head -> origin/gh/anijain2305/941/head 2025-12-04T09:20:54.2162810Z * [new branch] gh/anijain2305/941/orig -> origin/gh/anijain2305/941/orig 2025-12-04T09:20:54.2165681Z * [new branch] gh/anijain2305/942/base -> origin/gh/anijain2305/942/base 2025-12-04T09:20:54.2166151Z * [new branch] gh/anijain2305/942/head -> origin/gh/anijain2305/942/head 2025-12-04T09:20:54.2166559Z * [new branch] gh/anijain2305/942/orig -> origin/gh/anijain2305/942/orig 2025-12-04T09:20:54.2166963Z * [new branch] gh/anijain2305/943/base -> origin/gh/anijain2305/943/base 2025-12-04T09:20:54.2167444Z * [new branch] gh/anijain2305/943/head -> origin/gh/anijain2305/943/head 2025-12-04T09:20:54.2167937Z * [new branch] gh/anijain2305/943/orig -> origin/gh/anijain2305/943/orig 2025-12-04T09:20:54.2170442Z * [new branch] gh/anijain2305/944/base -> origin/gh/anijain2305/944/base 2025-12-04T09:20:54.2171854Z * [new branch] gh/anijain2305/944/head -> origin/gh/anijain2305/944/head 2025-12-04T09:20:54.2172233Z * [new branch] gh/anijain2305/944/orig -> origin/gh/anijain2305/944/orig 2025-12-04T09:20:54.2172612Z * [new branch] gh/anijain2305/945/base -> origin/gh/anijain2305/945/base 2025-12-04T09:20:54.2174340Z * [new branch] gh/anijain2305/945/head -> origin/gh/anijain2305/945/head 2025-12-04T09:20:54.2175189Z * [new branch] gh/anijain2305/945/orig -> origin/gh/anijain2305/945/orig 2025-12-04T09:20:54.2175812Z * [new branch] gh/anijain2305/946/base -> origin/gh/anijain2305/946/base 2025-12-04T09:20:54.2176574Z * [new branch] gh/anijain2305/946/head -> origin/gh/anijain2305/946/head 2025-12-04T09:20:54.2177169Z * [new branch] gh/anijain2305/946/orig -> origin/gh/anijain2305/946/orig 2025-12-04T09:20:54.2177625Z * [new branch] gh/anijain2305/947/base -> origin/gh/anijain2305/947/base 2025-12-04T09:20:54.2178837Z * [new branch] gh/anijain2305/947/head -> origin/gh/anijain2305/947/head 2025-12-04T09:20:54.2179230Z * [new branch] gh/anijain2305/947/orig -> origin/gh/anijain2305/947/orig 2025-12-04T09:20:54.2184049Z * [new branch] gh/anijain2305/948/base -> origin/gh/anijain2305/948/base 2025-12-04T09:20:54.2187547Z * [new branch] gh/anijain2305/948/head -> origin/gh/anijain2305/948/head 2025-12-04T09:20:54.2187991Z * [new branch] gh/anijain2305/948/orig -> origin/gh/anijain2305/948/orig 2025-12-04T09:20:54.2188367Z * [new branch] gh/anijain2305/949/base -> origin/gh/anijain2305/949/base 2025-12-04T09:20:54.2188740Z * [new branch] gh/anijain2305/949/head -> origin/gh/anijain2305/949/head 2025-12-04T09:20:54.2189122Z * [new branch] gh/anijain2305/949/orig -> origin/gh/anijain2305/949/orig 2025-12-04T09:20:54.2189496Z * [new branch] gh/anijain2305/950/base -> origin/gh/anijain2305/950/base 2025-12-04T09:20:54.2189866Z * [new branch] gh/anijain2305/950/head -> origin/gh/anijain2305/950/head 2025-12-04T09:20:54.2190253Z * [new branch] gh/anijain2305/950/orig -> origin/gh/anijain2305/950/orig 2025-12-04T09:20:54.2190630Z * [new branch] gh/anijain2305/951/base -> origin/gh/anijain2305/951/base 2025-12-04T09:20:54.2191004Z * [new branch] gh/anijain2305/951/head -> origin/gh/anijain2305/951/head 2025-12-04T09:20:54.2191377Z * [new branch] gh/anijain2305/951/orig -> origin/gh/anijain2305/951/orig 2025-12-04T09:20:54.2191741Z * [new branch] gh/anijain2305/952/base -> origin/gh/anijain2305/952/base 2025-12-04T09:20:54.2192109Z * [new branch] gh/anijain2305/952/head -> origin/gh/anijain2305/952/head 2025-12-04T09:20:54.2192484Z * [new branch] gh/anijain2305/952/orig -> origin/gh/anijain2305/952/orig 2025-12-04T09:20:54.2195842Z * [new branch] gh/anijain2305/953/base -> origin/gh/anijain2305/953/base 2025-12-04T09:20:54.2196236Z * [new branch] gh/anijain2305/953/head -> origin/gh/anijain2305/953/head 2025-12-04T09:20:54.2196614Z * [new branch] gh/anijain2305/953/orig -> origin/gh/anijain2305/953/orig 2025-12-04T09:20:54.2196999Z * [new branch] gh/anijain2305/954/base -> origin/gh/anijain2305/954/base 2025-12-04T09:20:54.2197372Z * [new branch] gh/anijain2305/954/head -> origin/gh/anijain2305/954/head 2025-12-04T09:20:54.2197891Z * [new branch] gh/anijain2305/954/orig -> origin/gh/anijain2305/954/orig 2025-12-04T09:20:54.2198907Z * [new branch] gh/anijain2305/955/base -> origin/gh/anijain2305/955/base 2025-12-04T09:20:54.2199516Z * [new branch] gh/anijain2305/955/head -> origin/gh/anijain2305/955/head 2025-12-04T09:20:54.2200066Z * [new branch] gh/anijain2305/955/orig -> origin/gh/anijain2305/955/orig 2025-12-04T09:20:54.2201051Z * [new branch] gh/anijain2305/956/base -> origin/gh/anijain2305/956/base 2025-12-04T09:20:54.2201646Z * [new branch] gh/anijain2305/956/head -> origin/gh/anijain2305/956/head 2025-12-04T09:20:54.2202556Z * [new branch] gh/anijain2305/956/orig -> origin/gh/anijain2305/956/orig 2025-12-04T09:20:54.2204202Z * [new branch] gh/anijain2305/957/base -> origin/gh/anijain2305/957/base 2025-12-04T09:20:54.2204984Z * [new branch] gh/anijain2305/957/head -> origin/gh/anijain2305/957/head 2025-12-04T09:20:54.2205466Z * [new branch] gh/anijain2305/957/orig -> origin/gh/anijain2305/957/orig 2025-12-04T09:20:54.2206441Z * [new branch] gh/anijain2305/958/base -> origin/gh/anijain2305/958/base 2025-12-04T09:20:54.2207286Z * [new branch] gh/anijain2305/958/head -> origin/gh/anijain2305/958/head 2025-12-04T09:20:54.2207770Z * [new branch] gh/anijain2305/958/orig -> origin/gh/anijain2305/958/orig 2025-12-04T09:20:54.2210208Z * [new branch] gh/anijain2305/959/base -> origin/gh/anijain2305/959/base 2025-12-04T09:20:54.2214859Z * [new branch] gh/anijain2305/959/head -> origin/gh/anijain2305/959/head 2025-12-04T09:20:54.2215287Z * [new branch] gh/anijain2305/959/orig -> origin/gh/anijain2305/959/orig 2025-12-04T09:20:54.2215697Z * [new branch] gh/anijain2305/960/base -> origin/gh/anijain2305/960/base 2025-12-04T09:20:54.2216076Z * [new branch] gh/anijain2305/960/head -> origin/gh/anijain2305/960/head 2025-12-04T09:20:54.2216447Z * [new branch] gh/anijain2305/960/orig -> origin/gh/anijain2305/960/orig 2025-12-04T09:20:54.2217135Z * [new branch] gh/anijain2305/961/base -> origin/gh/anijain2305/961/base 2025-12-04T09:20:54.2217512Z * [new branch] gh/anijain2305/961/head -> origin/gh/anijain2305/961/head 2025-12-04T09:20:54.2217878Z * [new branch] gh/anijain2305/961/orig -> origin/gh/anijain2305/961/orig 2025-12-04T09:20:54.2218245Z * [new branch] gh/anijain2305/962/base -> origin/gh/anijain2305/962/base 2025-12-04T09:20:54.2218609Z * [new branch] gh/anijain2305/962/head -> origin/gh/anijain2305/962/head 2025-12-04T09:20:54.2218975Z * [new branch] gh/anijain2305/962/orig -> origin/gh/anijain2305/962/orig 2025-12-04T09:20:54.2224418Z * [new branch] gh/anijain2305/963/base -> origin/gh/anijain2305/963/base 2025-12-04T09:20:54.2224868Z * [new branch] gh/anijain2305/963/head -> origin/gh/anijain2305/963/head 2025-12-04T09:20:54.2225262Z * [new branch] gh/anijain2305/963/orig -> origin/gh/anijain2305/963/orig 2025-12-04T09:20:54.2225667Z * [new branch] gh/anijain2305/964/base -> origin/gh/anijain2305/964/base 2025-12-04T09:20:54.2226048Z * [new branch] gh/anijain2305/964/head -> origin/gh/anijain2305/964/head 2025-12-04T09:20:54.2226423Z * [new branch] gh/anijain2305/964/orig -> origin/gh/anijain2305/964/orig 2025-12-04T09:20:54.2226790Z * [new branch] gh/anijain2305/965/base -> origin/gh/anijain2305/965/base 2025-12-04T09:20:54.2227166Z * [new branch] gh/anijain2305/965/head -> origin/gh/anijain2305/965/head 2025-12-04T09:20:54.2227546Z * [new branch] gh/anijain2305/965/orig -> origin/gh/anijain2305/965/orig 2025-12-04T09:20:54.2227920Z * [new branch] gh/anijain2305/966/base -> origin/gh/anijain2305/966/base 2025-12-04T09:20:54.2228284Z * [new branch] gh/anijain2305/966/head -> origin/gh/anijain2305/966/head 2025-12-04T09:20:54.2228659Z * [new branch] gh/anijain2305/966/orig -> origin/gh/anijain2305/966/orig 2025-12-04T09:20:54.2229036Z * [new branch] gh/anijain2305/967/base -> origin/gh/anijain2305/967/base 2025-12-04T09:20:54.2229412Z * [new branch] gh/anijain2305/967/head -> origin/gh/anijain2305/967/head 2025-12-04T09:20:54.2230653Z * [new branch] gh/anijain2305/967/orig -> origin/gh/anijain2305/967/orig 2025-12-04T09:20:54.2234166Z * [new branch] gh/anijain2305/968/base -> origin/gh/anijain2305/968/base 2025-12-04T09:20:54.2234598Z * [new branch] gh/anijain2305/968/head -> origin/gh/anijain2305/968/head 2025-12-04T09:20:54.2235279Z * [new branch] gh/anijain2305/968/orig -> origin/gh/anijain2305/968/orig 2025-12-04T09:20:54.2235884Z * [new branch] gh/anijain2305/969/base -> origin/gh/anijain2305/969/base 2025-12-04T09:20:54.2236612Z * [new branch] gh/anijain2305/969/head -> origin/gh/anijain2305/969/head 2025-12-04T09:20:54.2238185Z * [new branch] gh/anijain2305/969/orig -> origin/gh/anijain2305/969/orig 2025-12-04T09:20:54.2238592Z * [new branch] gh/anijain2305/970/base -> origin/gh/anijain2305/970/base 2025-12-04T09:20:54.2239017Z * [new branch] gh/anijain2305/970/head -> origin/gh/anijain2305/970/head 2025-12-04T09:20:54.2239776Z * [new branch] gh/anijain2305/970/orig -> origin/gh/anijain2305/970/orig 2025-12-04T09:20:54.2241326Z * [new branch] gh/anjali411/216/base -> origin/gh/anjali411/216/base 2025-12-04T09:20:54.2242042Z * [new branch] gh/anjali411/216/head -> origin/gh/anjali411/216/head 2025-12-04T09:20:54.2242716Z * [new branch] gh/anjali411/216/orig -> origin/gh/anjali411/216/orig 2025-12-04T09:20:54.2244828Z * [new branch] gh/anshul-si/1/base -> origin/gh/anshul-si/1/base 2025-12-04T09:20:54.2245566Z * [new branch] gh/anshul-si/1/head -> origin/gh/anshul-si/1/head 2025-12-04T09:20:54.2246168Z * [new branch] gh/anshul-si/2/base -> origin/gh/anshul-si/2/base 2025-12-04T09:20:54.2246726Z * [new branch] gh/anshul-si/2/head -> origin/gh/anshul-si/2/head 2025-12-04T09:20:54.2247191Z * [new branch] gh/anshul-si/3/base -> origin/gh/anshul-si/3/base 2025-12-04T09:20:54.2247907Z * [new branch] gh/anshul-si/3/head -> origin/gh/anshul-si/3/head 2025-12-04T09:20:54.2249172Z * [new branch] gh/anshul-si/4/base -> origin/gh/anshul-si/4/base 2025-12-04T09:20:54.2249544Z * [new branch] gh/anshul-si/4/head -> origin/gh/anshul-si/4/head 2025-12-04T09:20:54.2250506Z * [new branch] gh/anshul-si/5/base -> origin/gh/anshul-si/5/base 2025-12-04T09:20:54.2251109Z * [new branch] gh/anshul-si/5/head -> origin/gh/anshul-si/5/head 2025-12-04T09:20:54.2252561Z * [new branch] gh/anshul-si/53/base -> origin/gh/anshul-si/53/base 2025-12-04T09:20:54.2253278Z * [new branch] gh/anshul-si/53/head -> origin/gh/anshul-si/53/head 2025-12-04T09:20:54.2254892Z * [new branch] gh/anshul-si/58/base -> origin/gh/anshul-si/58/base 2025-12-04T09:20:54.2255279Z * [new branch] gh/anshul-si/58/head -> origin/gh/anshul-si/58/head 2025-12-04T09:20:54.2256799Z * [new branch] gh/anshul-si/66/base -> origin/gh/anshul-si/66/base 2025-12-04T09:20:54.2257166Z * [new branch] gh/anshul-si/66/head -> origin/gh/anshul-si/66/head 2025-12-04T09:20:54.2257567Z * [new branch] gh/anshul-si/66/orig -> origin/gh/anshul-si/66/orig 2025-12-04T09:20:54.2258152Z * [new branch] gh/anshul-si/67/base -> origin/gh/anshul-si/67/base 2025-12-04T09:20:54.2258763Z * [new branch] gh/anshul-si/67/head -> origin/gh/anshul-si/67/head 2025-12-04T09:20:54.2259473Z * [new branch] gh/anshul-si/67/orig -> origin/gh/anshul-si/67/orig 2025-12-04T09:20:54.2260848Z * [new branch] gh/anshul-si/68/base -> origin/gh/anshul-si/68/base 2025-12-04T09:20:54.2261641Z * [new branch] gh/anshul-si/68/head -> origin/gh/anshul-si/68/head 2025-12-04T09:20:54.2262096Z * [new branch] gh/anshul-si/68/orig -> origin/gh/anshul-si/68/orig 2025-12-04T09:20:54.2263902Z * [new branch] gh/anshul-si/69/base -> origin/gh/anshul-si/69/base 2025-12-04T09:20:54.2264308Z * [new branch] gh/anshul-si/69/head -> origin/gh/anshul-si/69/head 2025-12-04T09:20:54.2266104Z * [new branch] gh/anshul-si/69/orig -> origin/gh/anshul-si/69/orig 2025-12-04T09:20:54.2266637Z * [new branch] gh/anshul-si/70/base -> origin/gh/anshul-si/70/base 2025-12-04T09:20:54.2267260Z * [new branch] gh/anshul-si/70/head -> origin/gh/anshul-si/70/head 2025-12-04T09:20:54.2268123Z * [new branch] gh/anshul-si/70/orig -> origin/gh/anshul-si/70/orig 2025-12-04T09:20:54.2268739Z * [new branch] gh/anshul-si/71/base -> origin/gh/anshul-si/71/base 2025-12-04T09:20:54.2269411Z * [new branch] gh/anshul-si/71/head -> origin/gh/anshul-si/71/head 2025-12-04T09:20:54.2270294Z * [new branch] gh/anshul-si/71/orig -> origin/gh/anshul-si/71/orig 2025-12-04T09:20:54.2271249Z * [new branch] gh/anshul-si/72/base -> origin/gh/anshul-si/72/base 2025-12-04T09:20:54.2272903Z * [new branch] gh/anshul-si/72/head -> origin/gh/anshul-si/72/head 2025-12-04T09:20:54.2273305Z * [new branch] gh/anshul-si/72/orig -> origin/gh/anshul-si/72/orig 2025-12-04T09:20:54.2273662Z * [new branch] gh/anshul-si/73/base -> origin/gh/anshul-si/73/base 2025-12-04T09:20:54.2278715Z * [new branch] gh/anshul-si/73/head -> origin/gh/anshul-si/73/head 2025-12-04T09:20:54.2279180Z * [new branch] gh/anshul-si/73/orig -> origin/gh/anshul-si/73/orig 2025-12-04T09:20:54.2279555Z * [new branch] gh/aorenste/132/base -> origin/gh/aorenste/132/base 2025-12-04T09:20:54.2279918Z * [new branch] gh/aorenste/132/head -> origin/gh/aorenste/132/head 2025-12-04T09:20:54.2280279Z * [new branch] gh/aorenste/134/base -> origin/gh/aorenste/134/base 2025-12-04T09:20:54.2280662Z * [new branch] gh/aorenste/134/head -> origin/gh/aorenste/134/head 2025-12-04T09:20:54.2281040Z * [new branch] gh/aorenste/134/orig -> origin/gh/aorenste/134/orig 2025-12-04T09:20:54.2281398Z * [new branch] gh/aorenste/139/base -> origin/gh/aorenste/139/base 2025-12-04T09:20:54.2282011Z * [new branch] gh/aorenste/139/head -> origin/gh/aorenste/139/head 2025-12-04T09:20:54.2282719Z * [new branch] gh/aorenste/139/orig -> origin/gh/aorenste/139/orig 2025-12-04T09:20:54.2285906Z * [new branch] gh/aorenste/141/base -> origin/gh/aorenste/141/base 2025-12-04T09:20:54.2286275Z * [new branch] gh/aorenste/141/head -> origin/gh/aorenste/141/head 2025-12-04T09:20:54.2286652Z * [new branch] gh/aorenste/145/base -> origin/gh/aorenste/145/base 2025-12-04T09:20:54.2287006Z * [new branch] gh/aorenste/145/head -> origin/gh/aorenste/145/head 2025-12-04T09:20:54.2291089Z * [new branch] gh/aorenste/145/orig -> origin/gh/aorenste/145/orig 2025-12-04T09:20:54.2291570Z * [new branch] gh/aorenste/146/base -> origin/gh/aorenste/146/base 2025-12-04T09:20:54.2291955Z * [new branch] gh/aorenste/146/head -> origin/gh/aorenste/146/head 2025-12-04T09:20:54.2292307Z * [new branch] gh/aorenste/146/orig -> origin/gh/aorenste/146/orig 2025-12-04T09:20:54.2292748Z * [new branch] gh/aorenste/147/base -> origin/gh/aorenste/147/base 2025-12-04T09:20:54.2297219Z * [new branch] gh/aorenste/147/head -> origin/gh/aorenste/147/head 2025-12-04T09:20:54.2300564Z * [new branch] gh/aorenste/147/orig -> origin/gh/aorenste/147/orig 2025-12-04T09:20:54.2305218Z * [new branch] gh/aorenste/148/base -> origin/gh/aorenste/148/base 2025-12-04T09:20:54.2309497Z * [new branch] gh/aorenste/148/head -> origin/gh/aorenste/148/head 2025-12-04T09:20:54.2313435Z * [new branch] gh/aorenste/148/orig -> origin/gh/aorenste/148/orig 2025-12-04T09:20:54.2314990Z * [new branch] gh/aorenste/149/base -> origin/gh/aorenste/149/base 2025-12-04T09:20:54.2315405Z * [new branch] gh/aorenste/149/head -> origin/gh/aorenste/149/head 2025-12-04T09:20:54.2315780Z * [new branch] gh/aorenste/149/orig -> origin/gh/aorenste/149/orig 2025-12-04T09:20:54.2316216Z * [new branch] gh/aorenste/150/base -> origin/gh/aorenste/150/base 2025-12-04T09:20:54.2316587Z * [new branch] gh/aorenste/150/head -> origin/gh/aorenste/150/head 2025-12-04T09:20:54.2316950Z * [new branch] gh/aorenste/150/orig -> origin/gh/aorenste/150/orig 2025-12-04T09:20:54.2317306Z * [new branch] gh/aorenste/151/base -> origin/gh/aorenste/151/base 2025-12-04T09:20:54.2317669Z * [new branch] gh/aorenste/151/head -> origin/gh/aorenste/151/head 2025-12-04T09:20:54.2318030Z * [new branch] gh/aorenste/151/orig -> origin/gh/aorenste/151/orig 2025-12-04T09:20:54.2318397Z * [new branch] gh/aorenste/152/base -> origin/gh/aorenste/152/base 2025-12-04T09:20:54.2318757Z * [new branch] gh/aorenste/152/head -> origin/gh/aorenste/152/head 2025-12-04T09:20:54.2319127Z * [new branch] gh/aorenste/152/orig -> origin/gh/aorenste/152/orig 2025-12-04T09:20:54.2319507Z * [new branch] gh/aorenste/153/base -> origin/gh/aorenste/153/base 2025-12-04T09:20:54.2319862Z * [new branch] gh/aorenste/153/head -> origin/gh/aorenste/153/head 2025-12-04T09:20:54.2320223Z * [new branch] gh/aorenste/153/orig -> origin/gh/aorenste/153/orig 2025-12-04T09:20:54.2320586Z * [new branch] gh/aorenste/154/base -> origin/gh/aorenste/154/base 2025-12-04T09:20:54.2320972Z * [new branch] gh/aorenste/154/head -> origin/gh/aorenste/154/head 2025-12-04T09:20:54.2321336Z * [new branch] gh/aorenste/154/orig -> origin/gh/aorenste/154/orig 2025-12-04T09:20:54.2322044Z * [new branch] gh/aorenste/155/base -> origin/gh/aorenste/155/base 2025-12-04T09:20:54.2322435Z * [new branch] gh/aorenste/155/head -> origin/gh/aorenste/155/head 2025-12-04T09:20:54.2322817Z * [new branch] gh/aorenste/155/orig -> origin/gh/aorenste/155/orig 2025-12-04T09:20:54.2323184Z * [new branch] gh/aorenste/156/base -> origin/gh/aorenste/156/base 2025-12-04T09:20:54.2323549Z * [new branch] gh/aorenste/156/head -> origin/gh/aorenste/156/head 2025-12-04T09:20:54.2323913Z * [new branch] gh/aorenste/156/orig -> origin/gh/aorenste/156/orig 2025-12-04T09:20:54.2324275Z * [new branch] gh/aorenste/157/base -> origin/gh/aorenste/157/base 2025-12-04T09:20:54.2324633Z * [new branch] gh/aorenste/157/head -> origin/gh/aorenste/157/head 2025-12-04T09:20:54.2324989Z * [new branch] gh/aorenste/157/orig -> origin/gh/aorenste/157/orig 2025-12-04T09:20:54.2325350Z * [new branch] gh/aorenste/158/base -> origin/gh/aorenste/158/base 2025-12-04T09:20:54.2325697Z * [new branch] gh/aorenste/158/head -> origin/gh/aorenste/158/head 2025-12-04T09:20:54.2326066Z * [new branch] gh/aorenste/158/orig -> origin/gh/aorenste/158/orig 2025-12-04T09:20:54.2326421Z * [new branch] gh/aorenste/159/base -> origin/gh/aorenste/159/base 2025-12-04T09:20:54.2326782Z * [new branch] gh/aorenste/159/head -> origin/gh/aorenste/159/head 2025-12-04T09:20:54.2327140Z * [new branch] gh/aorenste/159/orig -> origin/gh/aorenste/159/orig 2025-12-04T09:20:54.2327535Z * [new branch] gh/avikchaudhuri/1/base -> origin/gh/avikchaudhuri/1/base 2025-12-04T09:20:54.2327924Z * [new branch] gh/avikchaudhuri/1/head -> origin/gh/avikchaudhuri/1/head 2025-12-04T09:20:54.2328390Z * [new branch] gh/avikchaudhuri/2/base -> origin/gh/avikchaudhuri/2/base 2025-12-04T09:20:54.2328767Z * [new branch] gh/avikchaudhuri/2/head -> origin/gh/avikchaudhuri/2/head 2025-12-04T09:20:54.2329157Z * [new branch] gh/avikchaudhuri/2/orig -> origin/gh/avikchaudhuri/2/orig 2025-12-04T09:20:54.2329591Z * [new branch] gh/bdhirsh/666/base -> origin/gh/bdhirsh/666/base 2025-12-04T09:20:54.2329954Z * [new branch] gh/bdhirsh/666/head -> origin/gh/bdhirsh/666/head 2025-12-04T09:20:54.2330506Z * [new branch] gh/bdhirsh/666/orig -> origin/gh/bdhirsh/666/orig 2025-12-04T09:20:54.2330874Z * [new branch] gh/bdhirsh/668/base -> origin/gh/bdhirsh/668/base 2025-12-04T09:20:54.2331277Z * [new branch] gh/bdhirsh/668/head -> origin/gh/bdhirsh/668/head 2025-12-04T09:20:54.2333211Z * [new branch] gh/bdhirsh/668/orig -> origin/gh/bdhirsh/668/orig 2025-12-04T09:20:54.2333890Z * [new branch] gh/bdhirsh/669/base -> origin/gh/bdhirsh/669/base 2025-12-04T09:20:54.2334326Z * [new branch] gh/bdhirsh/669/head -> origin/gh/bdhirsh/669/head 2025-12-04T09:20:54.2334717Z * [new branch] gh/bdhirsh/669/orig -> origin/gh/bdhirsh/669/orig 2025-12-04T09:20:54.2336267Z * [new branch] gh/bdhirsh/670/base -> origin/gh/bdhirsh/670/base 2025-12-04T09:20:54.2336784Z * [new branch] gh/bdhirsh/670/head -> origin/gh/bdhirsh/670/head 2025-12-04T09:20:54.2337562Z * [new branch] gh/bdhirsh/670/orig -> origin/gh/bdhirsh/670/orig 2025-12-04T09:20:54.2338827Z * [new branch] gh/bdhirsh/672/base -> origin/gh/bdhirsh/672/base 2025-12-04T09:20:54.2339244Z * [new branch] gh/bdhirsh/672/head -> origin/gh/bdhirsh/672/head 2025-12-04T09:20:54.2340078Z * [new branch] gh/bdhirsh/672/orig -> origin/gh/bdhirsh/672/orig 2025-12-04T09:20:54.2341546Z * [new branch] gh/bdhirsh/675/base -> origin/gh/bdhirsh/675/base 2025-12-04T09:20:54.2342238Z * [new branch] gh/bdhirsh/675/head -> origin/gh/bdhirsh/675/head 2025-12-04T09:20:54.2342973Z * [new branch] gh/bdhirsh/675/orig -> origin/gh/bdhirsh/675/orig 2025-12-04T09:20:54.2344204Z * [new branch] gh/bdhirsh/676/base -> origin/gh/bdhirsh/676/base 2025-12-04T09:20:54.2344789Z * [new branch] gh/bdhirsh/676/head -> origin/gh/bdhirsh/676/head 2025-12-04T09:20:54.2345572Z * [new branch] gh/bdhirsh/676/orig -> origin/gh/bdhirsh/676/orig 2025-12-04T09:20:54.2346748Z * [new branch] gh/bdhirsh/677/base -> origin/gh/bdhirsh/677/base 2025-12-04T09:20:54.2347897Z * [new branch] gh/bdhirsh/677/head -> origin/gh/bdhirsh/677/head 2025-12-04T09:20:54.2348369Z * [new branch] gh/bdhirsh/677/orig -> origin/gh/bdhirsh/677/orig 2025-12-04T09:20:54.2349885Z * [new branch] gh/bdhirsh/678/base -> origin/gh/bdhirsh/678/base 2025-12-04T09:20:54.2350672Z * [new branch] gh/bdhirsh/678/head -> origin/gh/bdhirsh/678/head 2025-12-04T09:20:54.2351250Z * [new branch] gh/bdhirsh/678/orig -> origin/gh/bdhirsh/678/orig 2025-12-04T09:20:54.2353264Z * [new branch] gh/bdhirsh/679/base -> origin/gh/bdhirsh/679/base 2025-12-04T09:20:54.2353765Z * [new branch] gh/bdhirsh/679/head -> origin/gh/bdhirsh/679/head 2025-12-04T09:20:54.2353950Z * [new branch] gh/bdhirsh/679/orig -> origin/gh/bdhirsh/679/orig 2025-12-04T09:20:54.2355082Z * [new branch] gh/bdhirsh/680/base -> origin/gh/bdhirsh/680/base 2025-12-04T09:20:54.2356712Z * [new branch] gh/bdhirsh/680/head -> origin/gh/bdhirsh/680/head 2025-12-04T09:20:54.2357015Z * [new branch] gh/bdhirsh/680/orig -> origin/gh/bdhirsh/680/orig 2025-12-04T09:20:54.2357170Z * [new branch] gh/bdhirsh/681/base -> origin/gh/bdhirsh/681/base 2025-12-04T09:20:54.2359575Z * [new branch] gh/bdhirsh/681/head -> origin/gh/bdhirsh/681/head 2025-12-04T09:20:54.2360272Z * [new branch] gh/bdhirsh/681/orig -> origin/gh/bdhirsh/681/orig 2025-12-04T09:20:54.2360476Z * [new branch] gh/benjaminglass1/101/base -> origin/gh/benjaminglass1/101/base 2025-12-04T09:20:54.2360685Z * [new branch] gh/benjaminglass1/101/head -> origin/gh/benjaminglass1/101/head 2025-12-04T09:20:54.2362227Z * [new branch] gh/benjaminglass1/101/orig -> origin/gh/benjaminglass1/101/orig 2025-12-04T09:20:54.2362531Z * [new branch] gh/benjaminglass1/102/base -> origin/gh/benjaminglass1/102/base 2025-12-04T09:20:54.2363541Z * [new branch] gh/benjaminglass1/102/head -> origin/gh/benjaminglass1/102/head 2025-12-04T09:20:54.2363948Z * [new branch] gh/benjaminglass1/102/orig -> origin/gh/benjaminglass1/102/orig 2025-12-04T09:20:54.2366033Z * [new branch] gh/benjaminglass1/106/base -> origin/gh/benjaminglass1/106/base 2025-12-04T09:20:54.2366407Z * [new branch] gh/benjaminglass1/106/head -> origin/gh/benjaminglass1/106/head 2025-12-04T09:20:54.2367640Z * [new branch] gh/benjaminglass1/106/orig -> origin/gh/benjaminglass1/106/orig 2025-12-04T09:20:54.2367850Z * [new branch] gh/benjaminglass1/107/base -> origin/gh/benjaminglass1/107/base 2025-12-04T09:20:54.2373270Z * [new branch] gh/benjaminglass1/107/head -> origin/gh/benjaminglass1/107/head 2025-12-04T09:20:54.2377261Z * [new branch] gh/benjaminglass1/107/orig -> origin/gh/benjaminglass1/107/orig 2025-12-04T09:20:54.2380762Z * [new branch] gh/benjaminglass1/108/base -> origin/gh/benjaminglass1/108/base 2025-12-04T09:20:54.2385348Z * [new branch] gh/benjaminglass1/108/head -> origin/gh/benjaminglass1/108/head 2025-12-04T09:20:54.2389136Z * [new branch] gh/benjaminglass1/108/orig -> origin/gh/benjaminglass1/108/orig 2025-12-04T09:20:54.2393276Z * [new branch] gh/benjaminglass1/109/base -> origin/gh/benjaminglass1/109/base 2025-12-04T09:20:54.2393502Z * [new branch] gh/benjaminglass1/109/head -> origin/gh/benjaminglass1/109/head 2025-12-04T09:20:54.2393736Z * [new branch] gh/benjaminglass1/109/orig -> origin/gh/benjaminglass1/109/orig 2025-12-04T09:20:54.2393936Z * [new branch] gh/benjaminglass1/97/base -> origin/gh/benjaminglass1/97/base 2025-12-04T09:20:54.2394114Z * [new branch] gh/benjaminglass1/97/head -> origin/gh/benjaminglass1/97/head 2025-12-04T09:20:54.2394282Z * [new branch] gh/benjaminglass1/97/orig -> origin/gh/benjaminglass1/97/orig 2025-12-04T09:20:54.2394463Z * [new branch] gh/bobrenjc93/570/base -> origin/gh/bobrenjc93/570/base 2025-12-04T09:20:54.2394622Z * [new branch] gh/bobrenjc93/570/head -> origin/gh/bobrenjc93/570/head 2025-12-04T09:20:54.2394779Z * [new branch] gh/bobrenjc93/570/orig -> origin/gh/bobrenjc93/570/orig 2025-12-04T09:20:54.2394941Z * [new branch] gh/bobrenjc93/604/base -> origin/gh/bobrenjc93/604/base 2025-12-04T09:20:54.2395095Z * [new branch] gh/bobrenjc93/604/head -> origin/gh/bobrenjc93/604/head 2025-12-04T09:20:54.2395249Z * [new branch] gh/bobrenjc93/604/orig -> origin/gh/bobrenjc93/604/orig 2025-12-04T09:20:54.2395413Z * [new branch] gh/bobrenjc93/638/base -> origin/gh/bobrenjc93/638/base 2025-12-04T09:20:54.2395568Z * [new branch] gh/bobrenjc93/638/head -> origin/gh/bobrenjc93/638/head 2025-12-04T09:20:54.2395725Z * [new branch] gh/bobrenjc93/638/orig -> origin/gh/bobrenjc93/638/orig 2025-12-04T09:20:54.2396027Z * [new branch] gh/bobrenjc93/653/base -> origin/gh/bobrenjc93/653/base 2025-12-04T09:20:54.2396179Z * [new branch] gh/bobrenjc93/653/head -> origin/gh/bobrenjc93/653/head 2025-12-04T09:20:54.2396340Z * [new branch] gh/bobrenjc93/653/orig -> origin/gh/bobrenjc93/653/orig 2025-12-04T09:20:54.2396577Z * [new branch] gh/bobrenjc93/654/base -> origin/gh/bobrenjc93/654/base 2025-12-04T09:20:54.2396736Z * [new branch] gh/bobrenjc93/654/head -> origin/gh/bobrenjc93/654/head 2025-12-04T09:20:54.2396888Z * [new branch] gh/bobrenjc93/654/orig -> origin/gh/bobrenjc93/654/orig 2025-12-04T09:20:54.2397034Z * [new branch] gh/bobrenjc93/657/base -> origin/gh/bobrenjc93/657/base 2025-12-04T09:20:54.2397190Z * [new branch] gh/bobrenjc93/657/head -> origin/gh/bobrenjc93/657/head 2025-12-04T09:20:54.2397343Z * [new branch] gh/bobrenjc93/657/orig -> origin/gh/bobrenjc93/657/orig 2025-12-04T09:20:54.2397496Z * [new branch] gh/bobrenjc93/672/base -> origin/gh/bobrenjc93/672/base 2025-12-04T09:20:54.2397643Z * [new branch] gh/bobrenjc93/672/head -> origin/gh/bobrenjc93/672/head 2025-12-04T09:20:54.2397794Z * [new branch] gh/bobrenjc93/672/orig -> origin/gh/bobrenjc93/672/orig 2025-12-04T09:20:54.2398118Z * [new branch] gh/bobrenjc93/679/base -> origin/gh/bobrenjc93/679/base 2025-12-04T09:20:54.2398272Z * [new branch] gh/bobrenjc93/679/head -> origin/gh/bobrenjc93/679/head 2025-12-04T09:20:54.2398430Z * [new branch] gh/bobrenjc93/679/orig -> origin/gh/bobrenjc93/679/orig 2025-12-04T09:20:54.2398578Z * [new branch] gh/bobrenjc93/680/base -> origin/gh/bobrenjc93/680/base 2025-12-04T09:20:54.2398730Z * [new branch] gh/bobrenjc93/680/head -> origin/gh/bobrenjc93/680/head 2025-12-04T09:20:54.2398982Z * [new branch] gh/bobrenjc93/680/orig -> origin/gh/bobrenjc93/680/orig 2025-12-04T09:20:54.2400244Z * [new branch] gh/bobrenjc93/681/base -> origin/gh/bobrenjc93/681/base 2025-12-04T09:20:54.2400580Z * [new branch] gh/bobrenjc93/681/head -> origin/gh/bobrenjc93/681/head 2025-12-04T09:20:54.2401914Z * [new branch] gh/bobrenjc93/681/orig -> origin/gh/bobrenjc93/681/orig 2025-12-04T09:20:54.2406540Z * [new branch] gh/bobrenjc93/682/base -> origin/gh/bobrenjc93/682/base 2025-12-04T09:20:54.2411480Z * [new branch] gh/bobrenjc93/682/head -> origin/gh/bobrenjc93/682/head 2025-12-04T09:20:54.2414115Z * [new branch] gh/bobrenjc93/682/orig -> origin/gh/bobrenjc93/682/orig 2025-12-04T09:20:54.2417411Z * [new branch] gh/bobrenjc93/683/base -> origin/gh/bobrenjc93/683/base 2025-12-04T09:20:54.2417617Z * [new branch] gh/bobrenjc93/683/head -> origin/gh/bobrenjc93/683/head 2025-12-04T09:20:54.2418209Z * [new branch] gh/bobrenjc93/683/orig -> origin/gh/bobrenjc93/683/orig 2025-12-04T09:20:54.2418405Z * [new branch] gh/bobrenjc93/684/base -> origin/gh/bobrenjc93/684/base 2025-12-04T09:20:54.2418575Z * [new branch] gh/bobrenjc93/684/head -> origin/gh/bobrenjc93/684/head 2025-12-04T09:20:54.2418747Z * [new branch] gh/bobrenjc93/684/orig -> origin/gh/bobrenjc93/684/orig 2025-12-04T09:20:54.2418914Z * [new branch] gh/bobrenjc93/685/base -> origin/gh/bobrenjc93/685/base 2025-12-04T09:20:54.2419070Z * [new branch] gh/bobrenjc93/685/head -> origin/gh/bobrenjc93/685/head 2025-12-04T09:20:54.2419228Z * [new branch] gh/bobrenjc93/685/orig -> origin/gh/bobrenjc93/685/orig 2025-12-04T09:20:54.2419398Z * [new branch] gh/bobrenjc93/686/base -> origin/gh/bobrenjc93/686/base 2025-12-04T09:20:54.2419689Z * [new branch] gh/bobrenjc93/686/head -> origin/gh/bobrenjc93/686/head 2025-12-04T09:20:54.2419852Z * [new branch] gh/bobrenjc93/686/orig -> origin/gh/bobrenjc93/686/orig 2025-12-04T09:20:54.2420009Z * [new branch] gh/bobrenjc93/687/base -> origin/gh/bobrenjc93/687/base 2025-12-04T09:20:54.2420215Z * [new branch] gh/bobrenjc93/687/head -> origin/gh/bobrenjc93/687/head 2025-12-04T09:20:54.2420374Z * [new branch] gh/bobrenjc93/687/orig -> origin/gh/bobrenjc93/687/orig 2025-12-04T09:20:54.2420790Z * [new branch] gh/bobrenjc93/688/base -> origin/gh/bobrenjc93/688/base 2025-12-04T09:20:54.2421065Z * [new branch] gh/bobrenjc93/688/head -> origin/gh/bobrenjc93/688/head 2025-12-04T09:20:54.2421231Z * [new branch] gh/bobrenjc93/688/orig -> origin/gh/bobrenjc93/688/orig 2025-12-04T09:20:54.2421389Z * [new branch] gh/bobrenjc93/689/base -> origin/gh/bobrenjc93/689/base 2025-12-04T09:20:54.2424738Z * [new branch] gh/bobrenjc93/689/head -> origin/gh/bobrenjc93/689/head 2025-12-04T09:20:54.2424917Z * [new branch] gh/bobrenjc93/689/orig -> origin/gh/bobrenjc93/689/orig 2025-12-04T09:20:54.2425079Z * [new branch] gh/bobrenjc93/690/base -> origin/gh/bobrenjc93/690/base 2025-12-04T09:20:54.2425249Z * [new branch] gh/bobrenjc93/690/head -> origin/gh/bobrenjc93/690/head 2025-12-04T09:20:54.2425806Z * [new branch] gh/bobrenjc93/690/orig -> origin/gh/bobrenjc93/690/orig 2025-12-04T09:20:54.2425974Z * [new branch] gh/bobrenjc93/691/base -> origin/gh/bobrenjc93/691/base 2025-12-04T09:20:54.2430655Z * [new branch] gh/bobrenjc93/691/head -> origin/gh/bobrenjc93/691/head 2025-12-04T09:20:54.2433618Z * [new branch] gh/bobrenjc93/691/orig -> origin/gh/bobrenjc93/691/orig 2025-12-04T09:20:54.2433842Z * [new branch] gh/bobrenjc93/692/base -> origin/gh/bobrenjc93/692/base 2025-12-04T09:20:54.2434013Z * [new branch] gh/bobrenjc93/692/head -> origin/gh/bobrenjc93/692/head 2025-12-04T09:20:54.2434185Z * [new branch] gh/bobrenjc93/692/orig -> origin/gh/bobrenjc93/692/orig 2025-12-04T09:20:54.2434391Z * [new branch] gh/bobrenjc93/693/base -> origin/gh/bobrenjc93/693/base 2025-12-04T09:20:54.2434557Z * [new branch] gh/bobrenjc93/693/head -> origin/gh/bobrenjc93/693/head 2025-12-04T09:20:54.2434766Z * [new branch] gh/bobrenjc93/693/orig -> origin/gh/bobrenjc93/693/orig 2025-12-04T09:20:54.2434930Z * [new branch] gh/bobrenjc93/694/base -> origin/gh/bobrenjc93/694/base 2025-12-04T09:20:54.2439950Z * [new branch] gh/bobrenjc93/694/head -> origin/gh/bobrenjc93/694/head 2025-12-04T09:20:54.2440130Z * [new branch] gh/bobrenjc93/694/orig -> origin/gh/bobrenjc93/694/orig 2025-12-04T09:20:54.2440307Z * [new branch] gh/bobrenjc93/695/base -> origin/gh/bobrenjc93/695/base 2025-12-04T09:20:54.2440467Z * [new branch] gh/bobrenjc93/695/head -> origin/gh/bobrenjc93/695/head 2025-12-04T09:20:54.2440619Z * [new branch] gh/bobrenjc93/695/orig -> origin/gh/bobrenjc93/695/orig 2025-12-04T09:20:54.2440804Z * [new branch] gh/c00w/23/base -> origin/gh/c00w/23/base 2025-12-04T09:20:54.2441191Z * [new branch] gh/c00w/23/head -> origin/gh/c00w/23/head 2025-12-04T09:20:54.2443492Z * [new branch] gh/c00w/53/base -> origin/gh/c00w/53/base 2025-12-04T09:20:54.2443650Z * [new branch] gh/c00w/53/head -> origin/gh/c00w/53/head 2025-12-04T09:20:54.2443777Z * [new branch] gh/c00w/53/orig -> origin/gh/c00w/53/orig 2025-12-04T09:20:54.2445457Z * [new branch] gh/c00w/54/base -> origin/gh/c00w/54/base 2025-12-04T09:20:54.2445879Z * [new branch] gh/c00w/54/head -> origin/gh/c00w/54/head 2025-12-04T09:20:54.2446016Z * [new branch] gh/c00w/54/orig -> origin/gh/c00w/54/orig 2025-12-04T09:20:54.2449922Z * [new branch] gh/c00w/56/base -> origin/gh/c00w/56/base 2025-12-04T09:20:54.2450659Z * [new branch] gh/c00w/56/head -> origin/gh/c00w/56/head 2025-12-04T09:20:54.2450839Z * [new branch] gh/c00w/56/orig -> origin/gh/c00w/56/orig 2025-12-04T09:20:54.2450980Z * [new branch] gh/c00w/57/base -> origin/gh/c00w/57/base 2025-12-04T09:20:54.2451121Z * [new branch] gh/c00w/57/head -> origin/gh/c00w/57/head 2025-12-04T09:20:54.2451285Z * [new branch] gh/c00w/57/orig -> origin/gh/c00w/57/orig 2025-12-04T09:20:54.2451756Z * [new branch] gh/c00w/58/base -> origin/gh/c00w/58/base 2025-12-04T09:20:54.2457068Z * [new branch] gh/c00w/58/head -> origin/gh/c00w/58/head 2025-12-04T09:20:54.2460526Z * [new branch] gh/c00w/58/orig -> origin/gh/c00w/58/orig 2025-12-04T09:20:54.2463861Z * [new branch] gh/clee2000/1/base -> origin/gh/clee2000/1/base 2025-12-04T09:20:54.2465839Z * [new branch] gh/clee2000/1/head -> origin/gh/clee2000/1/head 2025-12-04T09:20:54.2466007Z * [new branch] gh/clee2000/1/orig -> origin/gh/clee2000/1/orig 2025-12-04T09:20:54.2466504Z * [new branch] gh/coconutruben/1/base -> origin/gh/coconutruben/1/base 2025-12-04T09:20:54.2466706Z * [new branch] gh/coconutruben/1/head -> origin/gh/coconutruben/1/head 2025-12-04T09:20:54.2466881Z * [new branch] gh/coconutruben/55/base -> origin/gh/coconutruben/55/base 2025-12-04T09:20:54.2467043Z * [new branch] gh/coconutruben/55/head -> origin/gh/coconutruben/55/head 2025-12-04T09:20:54.2467223Z * [new branch] gh/coconutruben/55/orig -> origin/gh/coconutruben/55/orig 2025-12-04T09:20:54.2467380Z * [new branch] gh/coconutruben/57/base -> origin/gh/coconutruben/57/base 2025-12-04T09:20:54.2467539Z * [new branch] gh/coconutruben/57/head -> origin/gh/coconutruben/57/head 2025-12-04T09:20:54.2467706Z * [new branch] gh/coconutruben/57/orig -> origin/gh/coconutruben/57/orig 2025-12-04T09:20:54.2467861Z * [new branch] gh/coconutruben/70/base -> origin/gh/coconutruben/70/base 2025-12-04T09:20:54.2468031Z * [new branch] gh/coconutruben/70/head -> origin/gh/coconutruben/70/head 2025-12-04T09:20:54.2468185Z * [new branch] gh/coconutruben/70/orig -> origin/gh/coconutruben/70/orig 2025-12-04T09:20:54.2468341Z * [new branch] gh/coconutruben/71/base -> origin/gh/coconutruben/71/base 2025-12-04T09:20:54.2468506Z * [new branch] gh/coconutruben/71/head -> origin/gh/coconutruben/71/head 2025-12-04T09:20:54.2468662Z * [new branch] gh/coconutruben/71/orig -> origin/gh/coconutruben/71/orig 2025-12-04T09:20:54.2469348Z * [new branch] gh/coconutruben/72/base -> origin/gh/coconutruben/72/base 2025-12-04T09:20:54.2470319Z * [new branch] gh/coconutruben/72/head -> origin/gh/coconutruben/72/head 2025-12-04T09:20:54.2470718Z * [new branch] gh/coconutruben/72/orig -> origin/gh/coconutruben/72/orig 2025-12-04T09:20:54.2471956Z * [new branch] gh/coconutruben/73/base -> origin/gh/coconutruben/73/base 2025-12-04T09:20:54.2472596Z * [new branch] gh/coconutruben/73/head -> origin/gh/coconutruben/73/head 2025-12-04T09:20:54.2473819Z * [new branch] gh/coconutruben/73/orig -> origin/gh/coconutruben/73/orig 2025-12-04T09:20:54.2474205Z * [new branch] gh/coconutruben/74/base -> origin/gh/coconutruben/74/base 2025-12-04T09:20:54.2475285Z * [new branch] gh/coconutruben/74/head -> origin/gh/coconutruben/74/head 2025-12-04T09:20:54.2475845Z * [new branch] gh/coconutruben/74/orig -> origin/gh/coconutruben/74/orig 2025-12-04T09:20:54.2477566Z * [new branch] gh/coconutruben/79/base -> origin/gh/coconutruben/79/base 2025-12-04T09:20:54.2478006Z * [new branch] gh/coconutruben/79/head -> origin/gh/coconutruben/79/head 2025-12-04T09:20:54.2478655Z * [new branch] gh/coconutruben/79/orig -> origin/gh/coconutruben/79/orig 2025-12-04T09:20:54.2482706Z * [new branch] gh/coconutruben/80/base -> origin/gh/coconutruben/80/base 2025-12-04T09:20:54.2485341Z * [new branch] gh/coconutruben/80/head -> origin/gh/coconutruben/80/head 2025-12-04T09:20:54.2485537Z * [new branch] gh/coconutruben/80/orig -> origin/gh/coconutruben/80/orig 2025-12-04T09:20:54.2485699Z * [new branch] gh/coconutruben/82/base -> origin/gh/coconutruben/82/base 2025-12-04T09:20:54.2485877Z * [new branch] gh/coconutruben/82/head -> origin/gh/coconutruben/82/head 2025-12-04T09:20:54.2486043Z * [new branch] gh/coconutruben/82/orig -> origin/gh/coconutruben/82/orig 2025-12-04T09:20:54.2486203Z * [new branch] gh/coconutruben/83/base -> origin/gh/coconutruben/83/base 2025-12-04T09:20:54.2489586Z * [new branch] gh/coconutruben/83/head -> origin/gh/coconutruben/83/head 2025-12-04T09:20:54.2489750Z * [new branch] gh/coconutruben/83/orig -> origin/gh/coconutruben/83/orig 2025-12-04T09:20:54.2489910Z * [new branch] gh/coconutruben/84/base -> origin/gh/coconutruben/84/base 2025-12-04T09:20:54.2490072Z * [new branch] gh/coconutruben/84/head -> origin/gh/coconutruben/84/head 2025-12-04T09:20:54.2490276Z * [new branch] gh/coconutruben/84/orig -> origin/gh/coconutruben/84/orig 2025-12-04T09:20:54.2490458Z * [new branch] gh/coconutruben/85/base -> origin/gh/coconutruben/85/base 2025-12-04T09:20:54.2494538Z * [new branch] gh/coconutruben/85/head -> origin/gh/coconutruben/85/head 2025-12-04T09:20:54.2494895Z * [new branch] gh/coconutruben/85/orig -> origin/gh/coconutruben/85/orig 2025-12-04T09:20:54.2495099Z * [new branch] gh/coconutruben/86/base -> origin/gh/coconutruben/86/base 2025-12-04T09:20:54.2495310Z * [new branch] gh/coconutruben/86/head -> origin/gh/coconutruben/86/head 2025-12-04T09:20:54.2495493Z * [new branch] gh/coconutruben/86/orig -> origin/gh/coconutruben/86/orig 2025-12-04T09:20:54.2495678Z * [new branch] gh/colinchan15/1/base -> origin/gh/colinchan15/1/base 2025-12-04T09:20:54.2495849Z * [new branch] gh/colinchan15/1/head -> origin/gh/colinchan15/1/head 2025-12-04T09:20:54.2500782Z * [new branch] gh/colinchan15/2/base -> origin/gh/colinchan15/2/base 2025-12-04T09:20:54.2501109Z * [new branch] gh/colinchan15/2/head -> origin/gh/colinchan15/2/head 2025-12-04T09:20:54.2505935Z * [new branch] gh/colinchan15/3/base -> origin/gh/colinchan15/3/base 2025-12-04T09:20:54.2510968Z * [new branch] gh/colinchan15/3/head -> origin/gh/colinchan15/3/head 2025-12-04T09:20:54.2513155Z * [new branch] gh/colinchan15/6/base -> origin/gh/colinchan15/6/base 2025-12-04T09:20:54.2513464Z * [new branch] gh/colinchan15/6/head -> origin/gh/colinchan15/6/head 2025-12-04T09:20:54.2517012Z * [new branch] gh/d4l3k/1/base -> origin/gh/d4l3k/1/base 2025-12-04T09:20:54.2517324Z * [new branch] gh/d4l3k/1/head -> origin/gh/d4l3k/1/head 2025-12-04T09:20:54.2517495Z * [new branch] gh/d4l3k/2/base -> origin/gh/d4l3k/2/base 2025-12-04T09:20:54.2517634Z * [new branch] gh/d4l3k/2/head -> origin/gh/d4l3k/2/head 2025-12-04T09:20:54.2517934Z * [new branch] gh/d4l3k/2/orig -> origin/gh/d4l3k/2/orig 2025-12-04T09:20:54.2518069Z * [new branch] gh/d4l3k/3/base -> origin/gh/d4l3k/3/base 2025-12-04T09:20:54.2518223Z * [new branch] gh/d4l3k/3/head -> origin/gh/d4l3k/3/head 2025-12-04T09:20:54.2518415Z * [new branch] gh/d4l3k/3/orig -> origin/gh/d4l3k/3/orig 2025-12-04T09:20:54.2518556Z * [new branch] gh/d4l3k/4/base -> origin/gh/d4l3k/4/base 2025-12-04T09:20:54.2518685Z * [new branch] gh/d4l3k/4/head -> origin/gh/d4l3k/4/head 2025-12-04T09:20:54.2518811Z * [new branch] gh/d4l3k/4/orig -> origin/gh/d4l3k/4/orig 2025-12-04T09:20:54.2518948Z * [new branch] gh/d4l3k/5/base -> origin/gh/d4l3k/5/base 2025-12-04T09:20:54.2519076Z * [new branch] gh/d4l3k/5/orig -> origin/gh/d4l3k/5/orig 2025-12-04T09:20:54.2519271Z * [new branch] gh/davidberard98/392/base -> origin/gh/davidberard98/392/base 2025-12-04T09:20:54.2519437Z * [new branch] gh/davidberard98/392/head -> origin/gh/davidberard98/392/head 2025-12-04T09:20:54.2519599Z * [new branch] gh/davidberard98/392/orig -> origin/gh/davidberard98/392/orig 2025-12-04T09:20:54.2519770Z * [new branch] gh/davidberard98/399/base -> origin/gh/davidberard98/399/base 2025-12-04T09:20:54.2519928Z * [new branch] gh/davidberard98/399/head -> origin/gh/davidberard98/399/head 2025-12-04T09:20:54.2520096Z * [new branch] gh/davidberard98/399/orig -> origin/gh/davidberard98/399/orig 2025-12-04T09:20:54.2520261Z * [new branch] gh/desertfire/605/base -> origin/gh/desertfire/605/base 2025-12-04T09:20:54.2520420Z * [new branch] gh/desertfire/605/head -> origin/gh/desertfire/605/head 2025-12-04T09:20:54.2520581Z * [new branch] gh/desertfire/605/orig -> origin/gh/desertfire/605/orig 2025-12-04T09:20:54.2520731Z * [new branch] gh/desertfire/606/base -> origin/gh/desertfire/606/base 2025-12-04T09:20:54.2521027Z * [new branch] gh/desertfire/606/head -> origin/gh/desertfire/606/head 2025-12-04T09:20:54.2522115Z * [new branch] gh/desertfire/606/orig -> origin/gh/desertfire/606/orig 2025-12-04T09:20:54.2527654Z * [new branch] gh/desertfire/607/base -> origin/gh/desertfire/607/base 2025-12-04T09:20:54.2528188Z * [new branch] gh/desertfire/607/head -> origin/gh/desertfire/607/head 2025-12-04T09:20:54.2528350Z * [new branch] gh/desertfire/607/orig -> origin/gh/desertfire/607/orig 2025-12-04T09:20:54.2528517Z * [new branch] gh/desertfire/608/base -> origin/gh/desertfire/608/base 2025-12-04T09:20:54.2529586Z * [new branch] gh/desertfire/608/head -> origin/gh/desertfire/608/head 2025-12-04T09:20:54.2529774Z * [new branch] gh/desertfire/608/orig -> origin/gh/desertfire/608/orig 2025-12-04T09:20:54.2529935Z * [new branch] gh/desertfire/609/base -> origin/gh/desertfire/609/base 2025-12-04T09:20:54.2530212Z * [new branch] gh/desertfire/609/head -> origin/gh/desertfire/609/head 2025-12-04T09:20:54.2530395Z * [new branch] gh/desertfire/609/orig -> origin/gh/desertfire/609/orig 2025-12-04T09:20:54.2530555Z * [new branch] gh/desertfire/610/base -> origin/gh/desertfire/610/base 2025-12-04T09:20:54.2538832Z * [new branch] gh/desertfire/610/head -> origin/gh/desertfire/610/head 2025-12-04T09:20:54.2539238Z * [new branch] gh/desertfire/610/orig -> origin/gh/desertfire/610/orig 2025-12-04T09:20:54.2539405Z * [new branch] gh/desertfire/611/base -> origin/gh/desertfire/611/base 2025-12-04T09:20:54.2539817Z * [new branch] gh/desertfire/611/head -> origin/gh/desertfire/611/head 2025-12-04T09:20:54.2539995Z * [new branch] gh/desertfire/611/orig -> origin/gh/desertfire/611/orig 2025-12-04T09:20:54.2540161Z * [new branch] gh/desertfire/612/base -> origin/gh/desertfire/612/base 2025-12-04T09:20:54.2540389Z * [new branch] gh/desertfire/612/head -> origin/gh/desertfire/612/head 2025-12-04T09:20:54.2540567Z * [new branch] gh/desertfire/612/orig -> origin/gh/desertfire/612/orig 2025-12-04T09:20:54.2540737Z * [new branch] gh/desertfire/613/base -> origin/gh/desertfire/613/base 2025-12-04T09:20:54.2540893Z * [new branch] gh/desertfire/613/head -> origin/gh/desertfire/613/head 2025-12-04T09:20:54.2541213Z * [new branch] gh/desertfire/613/orig -> origin/gh/desertfire/613/orig 2025-12-04T09:20:54.2546828Z * [new branch] gh/desertfire/614/base -> origin/gh/desertfire/614/base 2025-12-04T09:20:54.2547221Z * [new branch] gh/desertfire/614/head -> origin/gh/desertfire/614/head 2025-12-04T09:20:54.2547438Z * [new branch] gh/desertfire/614/orig -> origin/gh/desertfire/614/orig 2025-12-04T09:20:54.2547606Z * [new branch] gh/desertfire/615/base -> origin/gh/desertfire/615/base 2025-12-04T09:20:54.2547783Z * [new branch] gh/desertfire/615/head -> origin/gh/desertfire/615/head 2025-12-04T09:20:54.2547941Z * [new branch] gh/desertfire/615/orig -> origin/gh/desertfire/615/orig 2025-12-04T09:20:54.2548098Z * [new branch] gh/desertfire/616/base -> origin/gh/desertfire/616/base 2025-12-04T09:20:54.2553912Z * [new branch] gh/desertfire/616/head -> origin/gh/desertfire/616/head 2025-12-04T09:20:54.2554287Z * [new branch] gh/desertfire/616/orig -> origin/gh/desertfire/616/orig 2025-12-04T09:20:54.2554481Z * [new branch] gh/desertfire/617/base -> origin/gh/desertfire/617/base 2025-12-04T09:20:54.2554665Z * [new branch] gh/desertfire/617/head -> origin/gh/desertfire/617/head 2025-12-04T09:20:54.2554957Z * [new branch] gh/desertfire/617/orig -> origin/gh/desertfire/617/orig 2025-12-04T09:20:54.2555501Z * [new branch] gh/dharakk/1/base -> origin/gh/dharakk/1/base 2025-12-04T09:20:54.2555694Z * [new branch] gh/dharakk/1/head -> origin/gh/dharakk/1/head 2025-12-04T09:20:54.2555882Z * [new branch] gh/drisspg/170/base -> origin/gh/drisspg/170/base 2025-12-04T09:20:54.2556033Z * [new branch] gh/drisspg/170/head -> origin/gh/drisspg/170/head 2025-12-04T09:20:54.2556176Z * [new branch] gh/drisspg/170/orig -> origin/gh/drisspg/170/orig 2025-12-04T09:20:54.2556329Z * [new branch] gh/drisspg/182/base -> origin/gh/drisspg/182/base 2025-12-04T09:20:54.2556484Z * [new branch] gh/drisspg/182/head -> origin/gh/drisspg/182/head 2025-12-04T09:20:54.2556643Z * [new branch] gh/drisspg/183/base -> origin/gh/drisspg/183/base 2025-12-04T09:20:54.2556782Z * [new branch] gh/drisspg/183/head -> origin/gh/drisspg/183/head 2025-12-04T09:20:54.2561288Z * [new branch] gh/drisspg/184/base -> origin/gh/drisspg/184/base 2025-12-04T09:20:54.2561496Z * [new branch] gh/drisspg/184/head -> origin/gh/drisspg/184/head 2025-12-04T09:20:54.2561660Z * [new branch] gh/drisspg/185/base -> origin/gh/drisspg/185/base 2025-12-04T09:20:54.2562055Z * [new branch] gh/drisspg/185/head -> origin/gh/drisspg/185/head 2025-12-04T09:20:54.2562209Z * [new branch] gh/drisspg/194/base -> origin/gh/drisspg/194/base 2025-12-04T09:20:54.2562354Z * [new branch] gh/drisspg/194/head -> origin/gh/drisspg/194/head 2025-12-04T09:20:54.2562663Z * [new branch] gh/drisspg/194/orig -> origin/gh/drisspg/194/orig 2025-12-04T09:20:54.2568645Z * [new branch] gh/drisspg/200/base -> origin/gh/drisspg/200/base 2025-12-04T09:20:54.2568957Z * [new branch] gh/drisspg/200/head -> origin/gh/drisspg/200/head 2025-12-04T09:20:54.2569360Z * [new branch] gh/drisspg/200/orig -> origin/gh/drisspg/200/orig 2025-12-04T09:20:54.2569542Z * [new branch] gh/drisspg/218/base -> origin/gh/drisspg/218/base 2025-12-04T09:20:54.2569691Z * [new branch] gh/drisspg/218/head -> origin/gh/drisspg/218/head 2025-12-04T09:20:54.2569848Z * [new branch] gh/drisspg/218/orig -> origin/gh/drisspg/218/orig 2025-12-04T09:20:54.2569994Z * [new branch] gh/drisspg/219/base -> origin/gh/drisspg/219/base 2025-12-04T09:20:54.2570145Z * [new branch] gh/drisspg/219/head -> origin/gh/drisspg/219/head 2025-12-04T09:20:54.2570304Z * [new branch] gh/drisspg/219/orig -> origin/gh/drisspg/219/orig 2025-12-04T09:20:54.2570464Z * [new branch] gh/drisspg/220/base -> origin/gh/drisspg/220/base 2025-12-04T09:20:54.2570622Z * [new branch] gh/drisspg/220/head -> origin/gh/drisspg/220/head 2025-12-04T09:20:54.2574634Z * [new branch] gh/drisspg/220/orig -> origin/gh/drisspg/220/orig 2025-12-04T09:20:54.2574842Z * [new branch] gh/drisspg/221/base -> origin/gh/drisspg/221/base 2025-12-04T09:20:54.2575006Z * [new branch] gh/drisspg/221/head -> origin/gh/drisspg/221/head 2025-12-04T09:20:54.2575155Z * [new branch] gh/drisspg/221/orig -> origin/gh/drisspg/221/orig 2025-12-04T09:20:54.2575308Z * [new branch] gh/drisspg/222/base -> origin/gh/drisspg/222/base 2025-12-04T09:20:54.2575455Z * [new branch] gh/drisspg/222/head -> origin/gh/drisspg/222/head 2025-12-04T09:20:54.2577747Z * [new branch] gh/drisspg/222/orig -> origin/gh/drisspg/222/orig 2025-12-04T09:20:54.2578326Z * [new branch] gh/drisspg/223/base -> origin/gh/drisspg/223/base 2025-12-04T09:20:54.2578523Z * [new branch] gh/drisspg/223/head -> origin/gh/drisspg/223/head 2025-12-04T09:20:54.2578715Z * [new branch] gh/drisspg/223/orig -> origin/gh/drisspg/223/orig 2025-12-04T09:20:54.2578863Z * [new branch] gh/drisspg/224/base -> origin/gh/drisspg/224/base 2025-12-04T09:20:54.2579048Z * [new branch] gh/drisspg/224/head -> origin/gh/drisspg/224/head 2025-12-04T09:20:54.2579597Z * [new branch] gh/drisspg/224/orig -> origin/gh/drisspg/224/orig 2025-12-04T09:20:54.2580970Z * [new branch] gh/drisspg/225/base -> origin/gh/drisspg/225/base 2025-12-04T09:20:54.2581169Z * [new branch] gh/drisspg/225/head -> origin/gh/drisspg/225/head 2025-12-04T09:20:54.2582363Z * [new branch] gh/drisspg/225/orig -> origin/gh/drisspg/225/orig 2025-12-04T09:20:54.2583664Z * [new branch] gh/drisspg/226/base -> origin/gh/drisspg/226/base 2025-12-04T09:20:54.2583828Z * [new branch] gh/drisspg/226/head -> origin/gh/drisspg/226/head 2025-12-04T09:20:54.2584847Z * [new branch] gh/drisspg/226/orig -> origin/gh/drisspg/226/orig 2025-12-04T09:20:54.2588234Z * [new branch] gh/drisspg/227/base -> origin/gh/drisspg/227/base 2025-12-04T09:20:54.2588948Z * [new branch] gh/drisspg/227/head -> origin/gh/drisspg/227/head 2025-12-04T09:20:54.2589188Z * [new branch] gh/drisspg/227/orig -> origin/gh/drisspg/227/orig 2025-12-04T09:20:54.2589348Z * [new branch] gh/drisspg/228/base -> origin/gh/drisspg/228/base 2025-12-04T09:20:54.2589494Z * [new branch] gh/drisspg/228/head -> origin/gh/drisspg/228/head 2025-12-04T09:20:54.2590052Z * [new branch] gh/drisspg/228/orig -> origin/gh/drisspg/228/orig 2025-12-04T09:20:54.2593745Z * [new branch] gh/drisspg/229/base -> origin/gh/drisspg/229/base 2025-12-04T09:20:54.2594251Z * [new branch] gh/drisspg/229/head -> origin/gh/drisspg/229/head 2025-12-04T09:20:54.2594527Z * [new branch] gh/drisspg/229/orig -> origin/gh/drisspg/229/orig 2025-12-04T09:20:54.2594678Z * [new branch] gh/drisspg/230/base -> origin/gh/drisspg/230/base 2025-12-04T09:20:54.2594832Z * [new branch] gh/drisspg/230/head -> origin/gh/drisspg/230/head 2025-12-04T09:20:54.2594991Z * [new branch] gh/drisspg/230/orig -> origin/gh/drisspg/230/orig 2025-12-04T09:20:54.2598477Z * [new branch] gh/dsjohns2/1/base -> origin/gh/dsjohns2/1/base 2025-12-04T09:20:54.2598659Z * [new branch] gh/dsjohns2/1/head -> origin/gh/dsjohns2/1/head 2025-12-04T09:20:54.2598846Z * [new branch] gh/dzmitry-huba/1/base -> origin/gh/dzmitry-huba/1/base 2025-12-04T09:20:54.2599425Z * [new branch] gh/dzmitry-huba/1/head -> origin/gh/dzmitry-huba/1/head 2025-12-04T09:20:54.2599834Z * [new branch] gh/dzmitry-huba/12/base -> origin/gh/dzmitry-huba/12/base 2025-12-04T09:20:54.2600860Z * [new branch] gh/dzmitry-huba/12/head -> origin/gh/dzmitry-huba/12/head 2025-12-04T09:20:54.2606096Z * [new branch] gh/dzmitry-huba/12/orig -> origin/gh/dzmitry-huba/12/orig 2025-12-04T09:20:54.2610664Z * [new branch] gh/dzmitry-huba/13/base -> origin/gh/dzmitry-huba/13/base 2025-12-04T09:20:54.2611624Z * [new branch] gh/dzmitry-huba/13/head -> origin/gh/dzmitry-huba/13/head 2025-12-04T09:20:54.2612175Z * [new branch] gh/dzmitry-huba/13/orig -> origin/gh/dzmitry-huba/13/orig 2025-12-04T09:20:54.2612392Z * [new branch] gh/dzmitry-huba/14/base -> origin/gh/dzmitry-huba/14/base 2025-12-04T09:20:54.2612557Z * [new branch] gh/dzmitry-huba/14/head -> origin/gh/dzmitry-huba/14/head 2025-12-04T09:20:54.2612719Z * [new branch] gh/dzmitry-huba/14/orig -> origin/gh/dzmitry-huba/14/orig 2025-12-04T09:20:54.2612885Z * [new branch] gh/dzmitry-huba/15/base -> origin/gh/dzmitry-huba/15/base 2025-12-04T09:20:54.2613045Z * [new branch] gh/dzmitry-huba/15/head -> origin/gh/dzmitry-huba/15/head 2025-12-04T09:20:54.2613199Z * [new branch] gh/dzmitry-huba/15/orig -> origin/gh/dzmitry-huba/15/orig 2025-12-04T09:20:54.2613352Z * [new branch] gh/dzmitry-huba/16/base -> origin/gh/dzmitry-huba/16/base 2025-12-04T09:20:54.2613511Z * [new branch] gh/dzmitry-huba/16/head -> origin/gh/dzmitry-huba/16/head 2025-12-04T09:20:54.2613665Z * [new branch] gh/dzmitry-huba/16/orig -> origin/gh/dzmitry-huba/16/orig 2025-12-04T09:20:54.2613822Z * [new branch] gh/dzmitry-huba/17/base -> origin/gh/dzmitry-huba/17/base 2025-12-04T09:20:54.2614026Z * [new branch] gh/dzmitry-huba/17/head -> origin/gh/dzmitry-huba/17/head 2025-12-04T09:20:54.2614201Z * [new branch] gh/dzmitry-huba/17/orig -> origin/gh/dzmitry-huba/17/orig 2025-12-04T09:20:54.2614389Z * [new branch] gh/dzmitry-huba/2/base -> origin/gh/dzmitry-huba/2/base 2025-12-04T09:20:54.2619634Z * [new branch] gh/dzmitry-huba/2/head -> origin/gh/dzmitry-huba/2/head 2025-12-04T09:20:54.2620816Z * [new branch] gh/dzmitry-huba/3/base -> origin/gh/dzmitry-huba/3/base 2025-12-04T09:20:54.2621319Z * [new branch] gh/dzmitry-huba/3/head -> origin/gh/dzmitry-huba/3/head 2025-12-04T09:20:54.2621521Z * [new branch] gh/eellison/808/base -> origin/gh/eellison/808/base 2025-12-04T09:20:54.2621831Z * [new branch] gh/eellison/808/head -> origin/gh/eellison/808/head 2025-12-04T09:20:54.2621979Z * [new branch] gh/eellison/808/orig -> origin/gh/eellison/808/orig 2025-12-04T09:20:54.2622138Z * [new branch] gh/eellison/822/base -> origin/gh/eellison/822/base 2025-12-04T09:20:54.2622356Z * [new branch] gh/eellison/822/head -> origin/gh/eellison/822/head 2025-12-04T09:20:54.2622508Z * [new branch] gh/eellison/822/orig -> origin/gh/eellison/822/orig 2025-12-04T09:20:54.2622927Z * [new branch] gh/eellison/823/base -> origin/gh/eellison/823/base 2025-12-04T09:20:54.2626334Z * [new branch] gh/eellison/823/head -> origin/gh/eellison/823/head 2025-12-04T09:20:54.2626534Z * [new branch] gh/eellison/823/orig -> origin/gh/eellison/823/orig 2025-12-04T09:20:54.2626689Z * [new branch] gh/eellison/862/base -> origin/gh/eellison/862/base 2025-12-04T09:20:54.2626865Z * [new branch] gh/eellison/862/head -> origin/gh/eellison/862/head 2025-12-04T09:20:54.2630794Z * [new branch] gh/eellison/862/orig -> origin/gh/eellison/862/orig 2025-12-04T09:20:54.2631409Z * [new branch] gh/eellison/863/base -> origin/gh/eellison/863/base 2025-12-04T09:20:54.2631613Z * [new branch] gh/eellison/863/head -> origin/gh/eellison/863/head 2025-12-04T09:20:54.2631888Z * [new branch] gh/eellison/863/orig -> origin/gh/eellison/863/orig 2025-12-04T09:20:54.2632062Z * [new branch] gh/eellison/864/base -> origin/gh/eellison/864/base 2025-12-04T09:20:54.2632219Z * [new branch] gh/eellison/864/head -> origin/gh/eellison/864/head 2025-12-04T09:20:54.2632384Z * [new branch] gh/eellison/864/orig -> origin/gh/eellison/864/orig 2025-12-04T09:20:54.2637406Z * [new branch] gh/eellison/865/base -> origin/gh/eellison/865/base 2025-12-04T09:20:54.2637613Z * [new branch] gh/eellison/865/head -> origin/gh/eellison/865/head 2025-12-04T09:20:54.2637777Z * [new branch] gh/eellison/865/orig -> origin/gh/eellison/865/orig 2025-12-04T09:20:54.2637924Z * [new branch] gh/eellison/866/base -> origin/gh/eellison/866/base 2025-12-04T09:20:54.2638102Z * [new branch] gh/eellison/866/head -> origin/gh/eellison/866/head 2025-12-04T09:20:54.2639329Z * [new branch] gh/eellison/866/orig -> origin/gh/eellison/866/orig 2025-12-04T09:20:54.2639624Z * [new branch] gh/eellison/867/base -> origin/gh/eellison/867/base 2025-12-04T09:20:54.2639794Z * [new branch] gh/eellison/867/head -> origin/gh/eellison/867/head 2025-12-04T09:20:54.2639948Z * [new branch] gh/eellison/867/orig -> origin/gh/eellison/867/orig 2025-12-04T09:20:54.2640122Z * [new branch] gh/eellison/868/base -> origin/gh/eellison/868/base 2025-12-04T09:20:54.2641999Z * [new branch] gh/eellison/868/head -> origin/gh/eellison/868/head 2025-12-04T09:20:54.2642228Z * [new branch] gh/eellison/868/orig -> origin/gh/eellison/868/orig 2025-12-04T09:20:54.2643186Z * [new branch] gh/eellison/869/base -> origin/gh/eellison/869/base 2025-12-04T09:20:54.2651085Z * [new branch] gh/eellison/869/head -> origin/gh/eellison/869/head 2025-12-04T09:20:54.2651495Z * [new branch] gh/eellison/869/orig -> origin/gh/eellison/869/orig 2025-12-04T09:20:54.2651652Z * [new branch] gh/eellison/870/base -> origin/gh/eellison/870/base 2025-12-04T09:20:54.2651807Z * [new branch] gh/eellison/870/head -> origin/gh/eellison/870/head 2025-12-04T09:20:54.2652102Z * [new branch] gh/eellison/870/orig -> origin/gh/eellison/870/orig 2025-12-04T09:20:54.2652538Z * [new branch] gh/eellison/871/base -> origin/gh/eellison/871/base 2025-12-04T09:20:54.2652691Z * [new branch] gh/eellison/871/head -> origin/gh/eellison/871/head 2025-12-04T09:20:54.2652843Z * [new branch] gh/eellison/871/orig -> origin/gh/eellison/871/orig 2025-12-04T09:20:54.2653090Z * [new branch] gh/eellison/872/base -> origin/gh/eellison/872/base 2025-12-04T09:20:54.2653236Z * [new branch] gh/eellison/872/head -> origin/gh/eellison/872/head 2025-12-04T09:20:54.2653384Z * [new branch] gh/eellison/872/orig -> origin/gh/eellison/872/orig 2025-12-04T09:20:54.2657398Z * [new branch] gh/eellison/873/base -> origin/gh/eellison/873/base 2025-12-04T09:20:54.2657588Z * [new branch] gh/eellison/873/head -> origin/gh/eellison/873/head 2025-12-04T09:20:54.2657757Z * [new branch] gh/eellison/873/orig -> origin/gh/eellison/873/orig 2025-12-04T09:20:54.2657921Z * [new branch] gh/eellison/874/base -> origin/gh/eellison/874/base 2025-12-04T09:20:54.2658087Z * [new branch] gh/eellison/874/head -> origin/gh/eellison/874/head 2025-12-04T09:20:54.2658238Z * [new branch] gh/eellison/874/orig -> origin/gh/eellison/874/orig 2025-12-04T09:20:54.2661932Z * [new branch] gh/eellison/875/base -> origin/gh/eellison/875/base 2025-12-04T09:20:54.2662143Z * [new branch] gh/eellison/875/head -> origin/gh/eellison/875/head 2025-12-04T09:20:54.2662296Z * [new branch] gh/eellison/875/orig -> origin/gh/eellison/875/orig 2025-12-04T09:20:54.2662452Z * [new branch] gh/eellison/876/base -> origin/gh/eellison/876/base 2025-12-04T09:20:54.2662599Z * [new branch] gh/eellison/876/head -> origin/gh/eellison/876/head 2025-12-04T09:20:54.2662747Z * [new branch] gh/eellison/876/orig -> origin/gh/eellison/876/orig 2025-12-04T09:20:54.2667138Z * [new branch] gh/eellison/877/base -> origin/gh/eellison/877/base 2025-12-04T09:20:54.2667711Z * [new branch] gh/eellison/877/head -> origin/gh/eellison/877/head 2025-12-04T09:20:54.2667908Z * [new branch] gh/eellison/877/orig -> origin/gh/eellison/877/orig 2025-12-04T09:20:54.2668077Z * [new branch] gh/eellison/878/base -> origin/gh/eellison/878/base 2025-12-04T09:20:54.2668226Z * [new branch] gh/eellison/878/head -> origin/gh/eellison/878/head 2025-12-04T09:20:54.2668720Z * [new branch] gh/eellison/878/orig -> origin/gh/eellison/878/orig 2025-12-04T09:20:54.2668867Z * [new branch] gh/eellison/879/base -> origin/gh/eellison/879/base 2025-12-04T09:20:54.2669019Z * [new branch] gh/eellison/879/head -> origin/gh/eellison/879/head 2025-12-04T09:20:54.2669171Z * [new branch] gh/eellison/879/orig -> origin/gh/eellison/879/orig 2025-12-04T09:20:54.2669530Z * [new branch] gh/eellison/880/base -> origin/gh/eellison/880/base 2025-12-04T09:20:54.2669795Z * [new branch] gh/eellison/880/head -> origin/gh/eellison/880/head 2025-12-04T09:20:54.2673900Z * [new branch] gh/eellison/880/orig -> origin/gh/eellison/880/orig 2025-12-04T09:20:54.2674096Z * [new branch] gh/eellison/881/base -> origin/gh/eellison/881/base 2025-12-04T09:20:54.2674921Z * [new branch] gh/eellison/881/head -> origin/gh/eellison/881/head 2025-12-04T09:20:54.2675439Z * [new branch] gh/eellison/881/orig -> origin/gh/eellison/881/orig 2025-12-04T09:20:54.2675632Z * [new branch] gh/eellison/882/base -> origin/gh/eellison/882/base 2025-12-04T09:20:54.2677987Z * [new branch] gh/eellison/882/head -> origin/gh/eellison/882/head 2025-12-04T09:20:54.2678872Z * [new branch] gh/eellison/882/orig -> origin/gh/eellison/882/orig 2025-12-04T09:20:54.2679067Z * [new branch] gh/eellison/883/base -> origin/gh/eellison/883/base 2025-12-04T09:20:54.2679226Z * [new branch] gh/eellison/883/head -> origin/gh/eellison/883/head 2025-12-04T09:20:54.2679573Z * [new branch] gh/eellison/883/orig -> origin/gh/eellison/883/orig 2025-12-04T09:20:54.2679745Z * [new branch] gh/eellison/884/base -> origin/gh/eellison/884/base 2025-12-04T09:20:54.2679907Z * [new branch] gh/eellison/884/head -> origin/gh/eellison/884/head 2025-12-04T09:20:54.2680391Z * [new branch] gh/eellison/884/orig -> origin/gh/eellison/884/orig 2025-12-04T09:20:54.2680549Z * [new branch] gh/etaf/147/base -> origin/gh/etaf/147/base 2025-12-04T09:20:54.2681101Z * [new branch] gh/etaf/147/head -> origin/gh/etaf/147/head 2025-12-04T09:20:54.2683679Z * [new branch] gh/etaf/154/base -> origin/gh/etaf/154/base 2025-12-04T09:20:54.2687023Z * [new branch] gh/etaf/154/head -> origin/gh/etaf/154/head 2025-12-04T09:20:54.2687380Z * [new branch] gh/etaf/154/orig -> origin/gh/etaf/154/orig 2025-12-04T09:20:54.2687559Z * [new branch] gh/etaf/156/base -> origin/gh/etaf/156/base 2025-12-04T09:20:54.2687702Z * [new branch] gh/etaf/156/head -> origin/gh/etaf/156/head 2025-12-04T09:20:54.2687851Z * [new branch] gh/etaf/156/orig -> origin/gh/etaf/156/orig 2025-12-04T09:20:54.2690579Z * [new branch] gh/etaf/157/base -> origin/gh/etaf/157/base 2025-12-04T09:20:54.2690928Z * [new branch] gh/etaf/157/head -> origin/gh/etaf/157/head 2025-12-04T09:20:54.2691081Z * [new branch] gh/etaf/157/orig -> origin/gh/etaf/157/orig 2025-12-04T09:20:54.2691247Z * [new branch] gh/etaf/158/base -> origin/gh/etaf/158/base 2025-12-04T09:20:54.2691382Z * [new branch] gh/etaf/158/head -> origin/gh/etaf/158/head 2025-12-04T09:20:54.2691525Z * [new branch] gh/etaf/158/orig -> origin/gh/etaf/158/orig 2025-12-04T09:20:54.2697535Z * [new branch] gh/etaf/159/base -> origin/gh/etaf/159/base 2025-12-04T09:20:54.2701996Z * [new branch] gh/etaf/159/head -> origin/gh/etaf/159/head 2025-12-04T09:20:54.2702500Z * [new branch] gh/etaf/159/orig -> origin/gh/etaf/159/orig 2025-12-04T09:20:54.2702656Z * [new branch] gh/etaf/160/base -> origin/gh/etaf/160/base 2025-12-04T09:20:54.2702802Z * [new branch] gh/etaf/160/head -> origin/gh/etaf/160/head 2025-12-04T09:20:54.2702944Z * [new branch] gh/etaf/160/orig -> origin/gh/etaf/160/orig 2025-12-04T09:20:54.2703115Z * [new branch] gh/etaf/161/base -> origin/gh/etaf/161/base 2025-12-04T09:20:54.2703260Z * [new branch] gh/etaf/161/head -> origin/gh/etaf/161/head 2025-12-04T09:20:54.2703405Z * [new branch] gh/etaf/161/orig -> origin/gh/etaf/161/orig 2025-12-04T09:20:54.2703559Z * [new branch] gh/etaf/166/base -> origin/gh/etaf/166/base 2025-12-04T09:20:54.2703692Z * [new branch] gh/etaf/166/head -> origin/gh/etaf/166/head 2025-12-04T09:20:54.2703836Z * [new branch] gh/etaf/166/orig -> origin/gh/etaf/166/orig 2025-12-04T09:20:54.2703986Z * [new branch] gh/etaf/167/base -> origin/gh/etaf/167/base 2025-12-04T09:20:54.2704282Z * [new branch] gh/etaf/167/head -> origin/gh/etaf/167/head 2025-12-04T09:20:54.2704764Z * [new branch] gh/etaf/167/orig -> origin/gh/etaf/167/orig 2025-12-04T09:20:54.2705141Z * [new branch] gh/etaf/168/base -> origin/gh/etaf/168/base 2025-12-04T09:20:54.2705666Z * [new branch] gh/etaf/168/head -> origin/gh/etaf/168/head 2025-12-04T09:20:54.2708439Z * [new branch] gh/etaf/168/orig -> origin/gh/etaf/168/orig 2025-12-04T09:20:54.2709115Z * [new branch] gh/etaf/172/base -> origin/gh/etaf/172/base 2025-12-04T09:20:54.2709412Z * [new branch] gh/etaf/172/head -> origin/gh/etaf/172/head 2025-12-04T09:20:54.2709573Z * [new branch] gh/etaf/172/orig -> origin/gh/etaf/172/orig 2025-12-04T09:20:54.2710959Z * [new branch] gh/etaf/173/base -> origin/gh/etaf/173/base 2025-12-04T09:20:54.2711193Z * [new branch] gh/etaf/173/head -> origin/gh/etaf/173/head 2025-12-04T09:20:54.2714271Z * [new branch] gh/etaf/173/orig -> origin/gh/etaf/173/orig 2025-12-04T09:20:54.2714512Z * [new branch] gh/etaf/174/base -> origin/gh/etaf/174/base 2025-12-04T09:20:54.2714658Z * [new branch] gh/etaf/174/head -> origin/gh/etaf/174/head 2025-12-04T09:20:54.2714978Z * [new branch] gh/etaf/175/base -> origin/gh/etaf/175/base 2025-12-04T09:20:54.2715284Z * [new branch] gh/etaf/175/head -> origin/gh/etaf/175/head 2025-12-04T09:20:54.2716802Z * [new branch] gh/etaf/175/orig -> origin/gh/etaf/175/orig 2025-12-04T09:20:54.2718888Z * [new branch] gh/etaf/176/base -> origin/gh/etaf/176/base 2025-12-04T09:20:54.2721419Z * [new branch] gh/etaf/176/head -> origin/gh/etaf/176/head 2025-12-04T09:20:54.2722124Z * [new branch] gh/etaf/176/orig -> origin/gh/etaf/176/orig 2025-12-04T09:20:54.2722316Z * [new branch] gh/etaf/177/base -> origin/gh/etaf/177/base 2025-12-04T09:20:54.2722476Z * [new branch] gh/etaf/177/head -> origin/gh/etaf/177/head 2025-12-04T09:20:54.2722617Z * [new branch] gh/etaf/177/orig -> origin/gh/etaf/177/orig 2025-12-04T09:20:54.2730711Z * [new branch] gh/etaf/178/base -> origin/gh/etaf/178/base 2025-12-04T09:20:54.2735720Z * [new branch] gh/etaf/178/head -> origin/gh/etaf/178/head 2025-12-04T09:20:54.2740058Z * [new branch] gh/etaf/178/orig -> origin/gh/etaf/178/orig 2025-12-04T09:20:54.2744466Z * [new branch] gh/etaf/179/base -> origin/gh/etaf/179/base 2025-12-04T09:20:54.2744648Z * [new branch] gh/etaf/179/head -> origin/gh/etaf/179/head 2025-12-04T09:20:54.2744807Z * [new branch] gh/etaf/179/orig -> origin/gh/etaf/179/orig 2025-12-04T09:20:54.2744956Z * [new branch] gh/etaf/180/base -> origin/gh/etaf/180/base 2025-12-04T09:20:54.2745137Z * [new branch] gh/etaf/180/head -> origin/gh/etaf/180/head 2025-12-04T09:20:54.2745282Z * [new branch] gh/etaf/180/orig -> origin/gh/etaf/180/orig 2025-12-04T09:20:54.2745476Z * [new branch] gh/exclamaforte/1/base -> origin/gh/exclamaforte/1/base 2025-12-04T09:20:54.2745646Z * [new branch] gh/exclamaforte/1/head -> origin/gh/exclamaforte/1/head 2025-12-04T09:20:54.2745813Z * [new branch] gh/exclamaforte/2/base -> origin/gh/exclamaforte/2/base 2025-12-04T09:20:54.2745976Z * [new branch] gh/exclamaforte/2/head -> origin/gh/exclamaforte/2/head 2025-12-04T09:20:54.2746139Z * [new branch] gh/exclamaforte/3/base -> origin/gh/exclamaforte/3/base 2025-12-04T09:20:54.2746301Z * [new branch] gh/exclamaforte/3/head -> origin/gh/exclamaforte/3/head 2025-12-04T09:20:54.2746471Z * [new branch] gh/exclamaforte/4/base -> origin/gh/exclamaforte/4/base 2025-12-04T09:20:54.2746870Z * [new branch] gh/exclamaforte/4/head -> origin/gh/exclamaforte/4/head 2025-12-04T09:20:54.2747035Z * [new branch] gh/ezyang/2374/base -> origin/gh/ezyang/2374/base 2025-12-04T09:20:54.2747181Z * [new branch] gh/ezyang/2374/head -> origin/gh/ezyang/2374/head 2025-12-04T09:20:54.2747388Z * [new branch] gh/ezyang/2374/orig -> origin/gh/ezyang/2374/orig 2025-12-04T09:20:54.2747548Z * [new branch] gh/ezyang/2973/base -> origin/gh/ezyang/2973/base 2025-12-04T09:20:54.2747689Z * [new branch] gh/ezyang/2973/head -> origin/gh/ezyang/2973/head 2025-12-04T09:20:54.2747841Z * [new branch] gh/ezyang/2973/orig -> origin/gh/ezyang/2973/orig 2025-12-04T09:20:54.2747985Z * [new branch] gh/ezyang/2974/base -> origin/gh/ezyang/2974/base 2025-12-04T09:20:54.2748128Z * [new branch] gh/ezyang/2974/head -> origin/gh/ezyang/2974/head 2025-12-04T09:20:54.2748284Z * [new branch] gh/ezyang/2974/orig -> origin/gh/ezyang/2974/orig 2025-12-04T09:20:54.2748426Z * [new branch] gh/ezyang/3131/base -> origin/gh/ezyang/3131/base 2025-12-04T09:20:54.2748574Z * [new branch] gh/ezyang/3131/head -> origin/gh/ezyang/3131/head 2025-12-04T09:20:54.2748719Z * [new branch] gh/ezyang/3131/orig -> origin/gh/ezyang/3131/orig 2025-12-04T09:20:54.2749103Z * [new branch] gh/ezyang/3139/base -> origin/gh/ezyang/3139/base 2025-12-04T09:20:54.2749256Z * [new branch] gh/ezyang/3139/head -> origin/gh/ezyang/3139/head 2025-12-04T09:20:54.2749397Z * [new branch] gh/ezyang/3139/orig -> origin/gh/ezyang/3139/orig 2025-12-04T09:20:54.2749557Z * [new branch] gh/ezyang/3140/base -> origin/gh/ezyang/3140/base 2025-12-04T09:20:54.2749695Z * [new branch] gh/ezyang/3140/head -> origin/gh/ezyang/3140/head 2025-12-04T09:20:54.2749976Z * [new branch] gh/ezyang/3140/orig -> origin/gh/ezyang/3140/orig 2025-12-04T09:20:54.2755236Z * [new branch] gh/ezyang/3143/base -> origin/gh/ezyang/3143/base 2025-12-04T09:20:54.2755588Z * [new branch] gh/ezyang/3143/head -> origin/gh/ezyang/3143/head 2025-12-04T09:20:54.2755776Z * [new branch] gh/ezyang/3143/orig -> origin/gh/ezyang/3143/orig 2025-12-04T09:20:54.2755933Z * [new branch] gh/ezyang/3144/base -> origin/gh/ezyang/3144/base 2025-12-04T09:20:54.2756077Z * [new branch] gh/ezyang/3144/head -> origin/gh/ezyang/3144/head 2025-12-04T09:20:54.2756373Z * [new branch] gh/ezyang/3144/orig -> origin/gh/ezyang/3144/orig 2025-12-04T09:20:54.2757753Z * [new branch] gh/ezyang/3167/base -> origin/gh/ezyang/3167/base 2025-12-04T09:20:54.2758007Z * [new branch] gh/ezyang/3167/head -> origin/gh/ezyang/3167/head 2025-12-04T09:20:54.2758207Z * [new branch] gh/ezyang/3167/orig -> origin/gh/ezyang/3167/orig 2025-12-04T09:20:54.2758355Z * [new branch] gh/ezyang/3173/base -> origin/gh/ezyang/3173/base 2025-12-04T09:20:54.2758632Z * [new branch] gh/ezyang/3173/head -> origin/gh/ezyang/3173/head 2025-12-04T09:20:54.2758798Z * [new branch] gh/ezyang/3173/orig -> origin/gh/ezyang/3173/orig 2025-12-04T09:20:54.2759265Z * [new branch] gh/ezyang/3175/base -> origin/gh/ezyang/3175/base 2025-12-04T09:20:54.2760221Z * [new branch] gh/ezyang/3175/head -> origin/gh/ezyang/3175/head 2025-12-04T09:20:54.2760624Z * [new branch] gh/ezyang/3175/orig -> origin/gh/ezyang/3175/orig 2025-12-04T09:20:54.2763163Z * [new branch] gh/ezyang/3182/base -> origin/gh/ezyang/3182/base 2025-12-04T09:20:54.2763505Z * [new branch] gh/ezyang/3182/head -> origin/gh/ezyang/3182/head 2025-12-04T09:20:54.2763668Z * [new branch] gh/ezyang/3182/orig -> origin/gh/ezyang/3182/orig 2025-12-04T09:20:54.2764581Z * [new branch] gh/ezyang/3185/base -> origin/gh/ezyang/3185/base 2025-12-04T09:20:54.2765107Z * [new branch] gh/ezyang/3185/head -> origin/gh/ezyang/3185/head 2025-12-04T09:20:54.2766071Z * [new branch] gh/ezyang/3185/orig -> origin/gh/ezyang/3185/orig 2025-12-04T09:20:54.2769966Z * [new branch] gh/ezyang/3189/base -> origin/gh/ezyang/3189/base 2025-12-04T09:20:54.2770160Z * [new branch] gh/ezyang/3189/head -> origin/gh/ezyang/3189/head 2025-12-04T09:20:54.2770708Z * [new branch] gh/ezyang/3189/orig -> origin/gh/ezyang/3189/orig 2025-12-04T09:20:54.2770891Z * [new branch] gh/ezyang/3191/base -> origin/gh/ezyang/3191/base 2025-12-04T09:20:54.2771055Z * [new branch] gh/ezyang/3191/head -> origin/gh/ezyang/3191/head 2025-12-04T09:20:54.2771198Z * [new branch] gh/ezyang/3191/orig -> origin/gh/ezyang/3191/orig 2025-12-04T09:20:54.2772820Z * [new branch] gh/ezyang/3192/base -> origin/gh/ezyang/3192/base 2025-12-04T09:20:54.2773469Z * [new branch] gh/ezyang/3192/head -> origin/gh/ezyang/3192/head 2025-12-04T09:20:54.2774170Z * [new branch] gh/ezyang/3192/orig -> origin/gh/ezyang/3192/orig 2025-12-04T09:20:54.2775214Z * [new branch] gh/ezyang/3193/base -> origin/gh/ezyang/3193/base 2025-12-04T09:20:54.2776817Z * [new branch] gh/ezyang/3193/head -> origin/gh/ezyang/3193/head 2025-12-04T09:20:54.2777155Z * [new branch] gh/ezyang/3193/orig -> origin/gh/ezyang/3193/orig 2025-12-04T09:20:54.2777396Z * [new branch] gh/ezyang/3194/base -> origin/gh/ezyang/3194/base 2025-12-04T09:20:54.2778278Z * [new branch] gh/ezyang/3194/head -> origin/gh/ezyang/3194/head 2025-12-04T09:20:54.2779642Z * [new branch] gh/ezyang/3194/orig -> origin/gh/ezyang/3194/orig 2025-12-04T09:20:54.2779911Z * [new branch] gh/ezyang/3195/base -> origin/gh/ezyang/3195/base 2025-12-04T09:20:54.2783969Z * [new branch] gh/ezyang/3195/head -> origin/gh/ezyang/3195/head 2025-12-04T09:20:54.2784160Z * [new branch] gh/ezyang/3195/orig -> origin/gh/ezyang/3195/orig 2025-12-04T09:20:54.2784307Z * [new branch] gh/ezyang/3196/base -> origin/gh/ezyang/3196/base 2025-12-04T09:20:54.2784454Z * [new branch] gh/ezyang/3196/head -> origin/gh/ezyang/3196/head 2025-12-04T09:20:54.2784606Z * [new branch] gh/ezyang/3196/orig -> origin/gh/ezyang/3196/orig 2025-12-04T09:20:54.2787123Z * [new branch] gh/ezyang/3197/base -> origin/gh/ezyang/3197/base 2025-12-04T09:20:54.2787305Z * [new branch] gh/ezyang/3197/head -> origin/gh/ezyang/3197/head 2025-12-04T09:20:54.2787542Z * [new branch] gh/ezyang/3197/orig -> origin/gh/ezyang/3197/orig 2025-12-04T09:20:54.2787808Z * [new branch] gh/ezyang/3198/base -> origin/gh/ezyang/3198/base 2025-12-04T09:20:54.2788135Z * [new branch] gh/ezyang/3198/head -> origin/gh/ezyang/3198/head 2025-12-04T09:20:54.2789937Z * [new branch] gh/ezyang/3198/orig -> origin/gh/ezyang/3198/orig 2025-12-04T09:20:54.2790288Z * [new branch] gh/ezyang/3199/base -> origin/gh/ezyang/3199/base 2025-12-04T09:20:54.2794621Z * [new branch] gh/ezyang/3199/head -> origin/gh/ezyang/3199/head 2025-12-04T09:20:54.2798559Z * [new branch] gh/ezyang/3199/orig -> origin/gh/ezyang/3199/orig 2025-12-04T09:20:54.2798754Z * [new branch] gh/ezyang/3200/base -> origin/gh/ezyang/3200/base 2025-12-04T09:20:54.2799088Z * [new branch] gh/ezyang/3200/head -> origin/gh/ezyang/3200/head 2025-12-04T09:20:54.2799235Z * [new branch] gh/ezyang/3200/orig -> origin/gh/ezyang/3200/orig 2025-12-04T09:20:54.2799391Z * [new branch] gh/ezyang/3201/base -> origin/gh/ezyang/3201/base 2025-12-04T09:20:54.2799598Z * [new branch] gh/ezyang/3201/head -> origin/gh/ezyang/3201/head 2025-12-04T09:20:54.2799760Z * [new branch] gh/ezyang/3201/orig -> origin/gh/ezyang/3201/orig 2025-12-04T09:20:54.2799904Z * [new branch] gh/ezyang/3202/base -> origin/gh/ezyang/3202/base 2025-12-04T09:20:54.2800045Z * [new branch] gh/ezyang/3202/head -> origin/gh/ezyang/3202/head 2025-12-04T09:20:54.2800192Z * [new branch] gh/ezyang/3202/orig -> origin/gh/ezyang/3202/orig 2025-12-04T09:20:54.2800337Z * [new branch] gh/ezyang/3203/base -> origin/gh/ezyang/3203/base 2025-12-04T09:20:54.2800490Z * [new branch] gh/ezyang/3203/head -> origin/gh/ezyang/3203/head 2025-12-04T09:20:54.2803736Z * [new branch] gh/ezyang/3203/orig -> origin/gh/ezyang/3203/orig 2025-12-04T09:20:54.2803892Z * [new branch] gh/ezyang/3204/base -> origin/gh/ezyang/3204/base 2025-12-04T09:20:54.2804044Z * [new branch] gh/ezyang/3204/head -> origin/gh/ezyang/3204/head 2025-12-04T09:20:54.2804186Z * [new branch] gh/ezyang/3204/orig -> origin/gh/ezyang/3204/orig 2025-12-04T09:20:54.2809776Z * [new branch] gh/ezyang/3205/base -> origin/gh/ezyang/3205/base 2025-12-04T09:20:54.2810006Z * [new branch] gh/ezyang/3205/head -> origin/gh/ezyang/3205/head 2025-12-04T09:20:54.2810757Z * [new branch] gh/ezyang/3205/orig -> origin/gh/ezyang/3205/orig 2025-12-04T09:20:54.2811073Z * [new branch] gh/ezyang/3206/base -> origin/gh/ezyang/3206/base 2025-12-04T09:20:54.2811220Z * [new branch] gh/ezyang/3206/head -> origin/gh/ezyang/3206/head 2025-12-04T09:20:54.2811358Z * [new branch] gh/ezyang/3206/orig -> origin/gh/ezyang/3206/orig 2025-12-04T09:20:54.2814887Z * [new branch] gh/ezyang/3207/base -> origin/gh/ezyang/3207/base 2025-12-04T09:20:54.2815148Z * [new branch] gh/ezyang/3207/head -> origin/gh/ezyang/3207/head 2025-12-04T09:20:54.2815310Z * [new branch] gh/ezyang/3207/orig -> origin/gh/ezyang/3207/orig 2025-12-04T09:20:54.2815455Z * [new branch] gh/ezyang/3208/base -> origin/gh/ezyang/3208/base 2025-12-04T09:20:54.2815610Z * [new branch] gh/ezyang/3208/head -> origin/gh/ezyang/3208/head 2025-12-04T09:20:54.2815754Z * [new branch] gh/ezyang/3208/orig -> origin/gh/ezyang/3208/orig 2025-12-04T09:20:54.2819400Z * [new branch] gh/ezyang/3209/base -> origin/gh/ezyang/3209/base 2025-12-04T09:20:54.2819578Z * [new branch] gh/ezyang/3209/head -> origin/gh/ezyang/3209/head 2025-12-04T09:20:54.2819767Z * [new branch] gh/ezyang/3209/orig -> origin/gh/ezyang/3209/orig 2025-12-04T09:20:54.2820057Z * [new branch] gh/fadara01/3/base -> origin/gh/fadara01/3/base 2025-12-04T09:20:54.2820225Z * [new branch] gh/fadara01/3/head -> origin/gh/fadara01/3/head 2025-12-04T09:20:54.2820379Z * [new branch] gh/fadara01/3/orig -> origin/gh/fadara01/3/orig 2025-12-04T09:20:54.2820620Z * [new branch] gh/fadara01/5/base -> origin/gh/fadara01/5/base 2025-12-04T09:20:54.2826230Z * [new branch] gh/fadara01/5/head -> origin/gh/fadara01/5/head 2025-12-04T09:20:54.2826419Z * [new branch] gh/fadara01/5/orig -> origin/gh/fadara01/5/orig 2025-12-04T09:20:54.2826730Z * [new branch] gh/fadara01/6/base -> origin/gh/fadara01/6/base 2025-12-04T09:20:54.2826875Z * [new branch] gh/fadara01/6/head -> origin/gh/fadara01/6/head 2025-12-04T09:20:54.2827030Z * [new branch] gh/fadara01/6/orig -> origin/gh/fadara01/6/orig 2025-12-04T09:20:54.2827251Z * [new branch] gh/fadara01/7/base -> origin/gh/fadara01/7/base 2025-12-04T09:20:54.2827397Z * [new branch] gh/fadara01/7/head -> origin/gh/fadara01/7/head 2025-12-04T09:20:54.2830406Z * [new branch] gh/fadara01/7/orig -> origin/gh/fadara01/7/orig 2025-12-04T09:20:54.2830596Z * [new branch] gh/fadara01/8/base -> origin/gh/fadara01/8/base 2025-12-04T09:20:54.2830761Z * [new branch] gh/fadara01/8/head -> origin/gh/fadara01/8/head 2025-12-04T09:20:54.2830904Z * [new branch] gh/fadara01/8/orig -> origin/gh/fadara01/8/orig 2025-12-04T09:20:54.2831073Z * [new branch] gh/fadara01/9/base -> origin/gh/fadara01/9/base 2025-12-04T09:20:54.2831214Z * [new branch] gh/fadara01/9/head -> origin/gh/fadara01/9/head 2025-12-04T09:20:54.2831356Z * [new branch] gh/fadara01/9/orig -> origin/gh/fadara01/9/orig 2025-12-04T09:20:54.2832719Z * [new branch] gh/fduwjj/182/base -> origin/gh/fduwjj/182/base 2025-12-04T09:20:54.2833176Z * [new branch] gh/fduwjj/182/head -> origin/gh/fduwjj/182/head 2025-12-04T09:20:54.2835098Z * [new branch] gh/fduwjj/182/orig -> origin/gh/fduwjj/182/orig 2025-12-04T09:20:54.2835997Z * [new branch] gh/fduwjj/211/base -> origin/gh/fduwjj/211/base 2025-12-04T09:20:54.2836661Z * [new branch] gh/fduwjj/211/head -> origin/gh/fduwjj/211/head 2025-12-04T09:20:54.2837499Z * [new branch] gh/fduwjj/211/orig -> origin/gh/fduwjj/211/orig 2025-12-04T09:20:54.2838530Z * [new branch] gh/fduwjj/212/base -> origin/gh/fduwjj/212/base 2025-12-04T09:20:54.2839031Z * [new branch] gh/fduwjj/212/head -> origin/gh/fduwjj/212/head 2025-12-04T09:20:54.2839992Z * [new branch] gh/fduwjj/212/orig -> origin/gh/fduwjj/212/orig 2025-12-04T09:20:54.2840984Z * [new branch] gh/fduwjj/213/base -> origin/gh/fduwjj/213/base 2025-12-04T09:20:54.2841587Z * [new branch] gh/fduwjj/213/head -> origin/gh/fduwjj/213/head 2025-12-04T09:20:54.2842451Z * [new branch] gh/fduwjj/213/orig -> origin/gh/fduwjj/213/orig 2025-12-04T09:20:54.2843906Z * [new branch] gh/fduwjj/226/base -> origin/gh/fduwjj/226/base 2025-12-04T09:20:54.2849052Z * [new branch] gh/fduwjj/226/head -> origin/gh/fduwjj/226/head 2025-12-04T09:20:54.2849416Z * [new branch] gh/fduwjj/226/orig -> origin/gh/fduwjj/226/orig 2025-12-04T09:20:54.2849855Z * [new branch] gh/fduwjj/229/base -> origin/gh/fduwjj/229/base 2025-12-04T09:20:54.2850006Z * [new branch] gh/fduwjj/229/head -> origin/gh/fduwjj/229/head 2025-12-04T09:20:54.2850155Z * [new branch] gh/fduwjj/229/orig -> origin/gh/fduwjj/229/orig 2025-12-04T09:20:54.2850322Z * [new branch] gh/fduwjj/233/base -> origin/gh/fduwjj/233/base 2025-12-04T09:20:54.2850557Z * [new branch] gh/fduwjj/233/head -> origin/gh/fduwjj/233/head 2025-12-04T09:20:54.2853611Z * [new branch] gh/fduwjj/233/orig -> origin/gh/fduwjj/233/orig 2025-12-04T09:20:54.2853939Z * [new branch] gh/fduwjj/234/base -> origin/gh/fduwjj/234/base 2025-12-04T09:20:54.2854109Z * [new branch] gh/fduwjj/234/head -> origin/gh/fduwjj/234/head 2025-12-04T09:20:54.2854267Z * [new branch] gh/fduwjj/234/orig -> origin/gh/fduwjj/234/orig 2025-12-04T09:20:54.2854668Z * [new branch] gh/fduwjj/235/base -> origin/gh/fduwjj/235/base 2025-12-04T09:20:54.2854819Z * [new branch] gh/fduwjj/235/head -> origin/gh/fduwjj/235/head 2025-12-04T09:20:54.2854964Z * [new branch] gh/fduwjj/235/orig -> origin/gh/fduwjj/235/orig 2025-12-04T09:20:54.2858853Z * [new branch] gh/fduwjj/236/base -> origin/gh/fduwjj/236/base 2025-12-04T09:20:54.2859000Z * [new branch] gh/fduwjj/236/head -> origin/gh/fduwjj/236/head 2025-12-04T09:20:54.2859148Z * [new branch] gh/fduwjj/236/orig -> origin/gh/fduwjj/236/orig 2025-12-04T09:20:54.2859282Z * [new branch] gh/fduwjj/237/base -> origin/gh/fduwjj/237/base 2025-12-04T09:20:54.2859431Z * [new branch] gh/fduwjj/237/head -> origin/gh/fduwjj/237/head 2025-12-04T09:20:54.2859569Z * [new branch] gh/fduwjj/237/orig -> origin/gh/fduwjj/237/orig 2025-12-04T09:20:54.2862539Z * [new branch] gh/fduwjj/238/base -> origin/gh/fduwjj/238/base 2025-12-04T09:20:54.2863081Z * [new branch] gh/fduwjj/238/head -> origin/gh/fduwjj/238/head 2025-12-04T09:20:54.2863360Z * [new branch] gh/fduwjj/238/orig -> origin/gh/fduwjj/238/orig 2025-12-04T09:20:54.2863535Z * [new branch] gh/fduwjj/239/base -> origin/gh/fduwjj/239/base 2025-12-04T09:20:54.2863678Z * [new branch] gh/fduwjj/239/head -> origin/gh/fduwjj/239/head 2025-12-04T09:20:54.2863835Z * [new branch] gh/fduwjj/239/orig -> origin/gh/fduwjj/239/orig 2025-12-04T09:20:54.2864876Z * [new branch] gh/fegin/332/base -> origin/gh/fegin/332/base 2025-12-04T09:20:54.2865393Z * [new branch] gh/fegin/332/head -> origin/gh/fegin/332/head 2025-12-04T09:20:54.2866318Z * [new branch] gh/fegin/332/orig -> origin/gh/fegin/332/orig 2025-12-04T09:20:54.2867416Z * [new branch] gh/fegin/333/base -> origin/gh/fegin/333/base 2025-12-04T09:20:54.2867874Z * [new branch] gh/fegin/333/head -> origin/gh/fegin/333/head 2025-12-04T09:20:54.2868985Z * [new branch] gh/fegin/333/orig -> origin/gh/fegin/333/orig 2025-12-04T09:20:54.2869883Z * [new branch] gh/fegin/334/base -> origin/gh/fegin/334/base 2025-12-04T09:20:54.2870371Z * [new branch] gh/fegin/334/head -> origin/gh/fegin/334/head 2025-12-04T09:20:54.2871446Z * [new branch] gh/fegin/334/orig -> origin/gh/fegin/334/orig 2025-12-04T09:20:54.2872389Z * [new branch] gh/fegin/335/base -> origin/gh/fegin/335/base 2025-12-04T09:20:54.2872950Z * [new branch] gh/fegin/335/head -> origin/gh/fegin/335/head 2025-12-04T09:20:54.2873732Z * [new branch] gh/fegin/335/orig -> origin/gh/fegin/335/orig 2025-12-04T09:20:54.2876783Z * [new branch] gh/fffrog/160/base -> origin/gh/fffrog/160/base 2025-12-04T09:20:54.2880116Z * [new branch] gh/fffrog/160/head -> origin/gh/fffrog/160/head 2025-12-04T09:20:54.2883875Z * [new branch] gh/fffrog/177/base -> origin/gh/fffrog/177/base 2025-12-04T09:20:54.2887957Z * [new branch] gh/fffrog/177/head -> origin/gh/fffrog/177/head 2025-12-04T09:20:54.2888116Z * [new branch] gh/fffrog/177/orig -> origin/gh/fffrog/177/orig 2025-12-04T09:20:54.2888282Z * [new branch] gh/fffrog/178/base -> origin/gh/fffrog/178/base 2025-12-04T09:20:54.2888482Z * [new branch] gh/fffrog/178/head -> origin/gh/fffrog/178/head 2025-12-04T09:20:54.2888631Z * [new branch] gh/fffrog/178/orig -> origin/gh/fffrog/178/orig 2025-12-04T09:20:54.2888927Z * [new branch] gh/fffrog/181/base -> origin/gh/fffrog/181/base 2025-12-04T09:20:54.2889163Z * [new branch] gh/fffrog/181/head -> origin/gh/fffrog/181/head 2025-12-04T09:20:54.2889324Z * [new branch] gh/fffrog/181/orig -> origin/gh/fffrog/181/orig 2025-12-04T09:20:54.2889517Z * [new branch] gh/fffrog/183/base -> origin/gh/fffrog/183/base 2025-12-04T09:20:54.2889670Z * [new branch] gh/fffrog/183/head -> origin/gh/fffrog/183/head 2025-12-04T09:20:54.2889809Z * [new branch] gh/fffrog/183/orig -> origin/gh/fffrog/183/orig 2025-12-04T09:20:54.2889953Z * [new branch] gh/fxdawnn/10/base -> origin/gh/fxdawnn/10/base 2025-12-04T09:20:54.2890102Z * [new branch] gh/fxdawnn/10/head -> origin/gh/fxdawnn/10/head 2025-12-04T09:20:54.2890241Z * [new branch] gh/fxdawnn/10/orig -> origin/gh/fxdawnn/10/orig 2025-12-04T09:20:54.2891645Z * [new branch] gh/fxdawnn/11/base -> origin/gh/fxdawnn/11/base 2025-12-04T09:20:54.2892217Z * [new branch] gh/fxdawnn/11/head -> origin/gh/fxdawnn/11/head 2025-12-04T09:20:54.2892464Z * [new branch] gh/fxdawnn/11/orig -> origin/gh/fxdawnn/11/orig 2025-12-04T09:20:54.2892619Z * [new branch] gh/fxdawnn/12/base -> origin/gh/fxdawnn/12/base 2025-12-04T09:20:54.2896132Z * [new branch] gh/fxdawnn/12/head -> origin/gh/fxdawnn/12/head 2025-12-04T09:20:54.2896459Z * [new branch] gh/fxdawnn/12/orig -> origin/gh/fxdawnn/12/orig 2025-12-04T09:20:54.2896618Z * [new branch] gh/fxdawnn/13/base -> origin/gh/fxdawnn/13/base 2025-12-04T09:20:54.2896772Z * [new branch] gh/fxdawnn/13/head -> origin/gh/fxdawnn/13/head 2025-12-04T09:20:54.2896915Z * [new branch] gh/fxdawnn/13/orig -> origin/gh/fxdawnn/13/orig 2025-12-04T09:20:54.2897081Z * [new branch] gh/fxdawnn/14/base -> origin/gh/fxdawnn/14/base 2025-12-04T09:20:54.2899972Z * [new branch] gh/fxdawnn/14/head -> origin/gh/fxdawnn/14/head 2025-12-04T09:20:54.2900130Z * [new branch] gh/fxdawnn/14/orig -> origin/gh/fxdawnn/14/orig 2025-12-04T09:20:54.2900285Z * [new branch] gh/fxdawnn/15/base -> origin/gh/fxdawnn/15/base 2025-12-04T09:20:54.2900425Z * [new branch] gh/fxdawnn/15/head -> origin/gh/fxdawnn/15/head 2025-12-04T09:20:54.2904172Z * [new branch] gh/fxdawnn/15/orig -> origin/gh/fxdawnn/15/orig 2025-12-04T09:20:54.2904375Z * [new branch] gh/fxdawnn/6/base -> origin/gh/fxdawnn/6/base 2025-12-04T09:20:54.2904526Z * [new branch] gh/fxdawnn/6/head -> origin/gh/fxdawnn/6/head 2025-12-04T09:20:54.2904663Z * [new branch] gh/fxdawnn/6/orig -> origin/gh/fxdawnn/6/orig 2025-12-04T09:20:54.2908508Z * [new branch] gh/fxdawnn/7/base -> origin/gh/fxdawnn/7/base 2025-12-04T09:20:54.2908692Z * [new branch] gh/fxdawnn/7/head -> origin/gh/fxdawnn/7/head 2025-12-04T09:20:54.2908846Z * [new branch] gh/fxdawnn/7/orig -> origin/gh/fxdawnn/7/orig 2025-12-04T09:20:54.2909014Z * [new branch] gh/fxdawnn/9/base -> origin/gh/fxdawnn/9/base 2025-12-04T09:20:54.2909157Z * [new branch] gh/fxdawnn/9/head -> origin/gh/fxdawnn/9/head 2025-12-04T09:20:54.2909294Z * [new branch] gh/fxdawnn/9/orig -> origin/gh/fxdawnn/9/orig 2025-12-04T09:20:54.2910779Z * [new branch] gh/galv/1/base -> origin/gh/galv/1/base 2025-12-04T09:20:54.2910941Z * [new branch] gh/galv/1/head -> origin/gh/galv/1/head 2025-12-04T09:20:54.2911083Z * [new branch] gh/galv/1/orig -> origin/gh/galv/1/orig 2025-12-04T09:20:54.2912416Z * [new branch] gh/galv/2/base -> origin/gh/galv/2/base 2025-12-04T09:20:54.2912895Z * [new branch] gh/galv/2/head -> origin/gh/galv/2/head 2025-12-04T09:20:54.2914494Z * [new branch] gh/galv/2/orig -> origin/gh/galv/2/orig 2025-12-04T09:20:54.2914814Z * [new branch] gh/galv/3/base -> origin/gh/galv/3/base 2025-12-04T09:20:54.2915143Z * [new branch] gh/galv/3/head -> origin/gh/galv/3/head 2025-12-04T09:20:54.2920097Z * [new branch] gh/galv/3/orig -> origin/gh/galv/3/orig 2025-12-04T09:20:54.2920381Z * [new branch] gh/guangyey/134/base -> origin/gh/guangyey/134/base 2025-12-04T09:20:54.2921067Z * [new branch] gh/guangyey/134/head -> origin/gh/guangyey/134/head 2025-12-04T09:20:54.2921257Z * [new branch] gh/guangyey/134/orig -> origin/gh/guangyey/134/orig 2025-12-04T09:20:54.2921426Z * [new branch] gh/guangyey/163/base -> origin/gh/guangyey/163/base 2025-12-04T09:20:54.2921583Z * [new branch] gh/guangyey/163/head -> origin/gh/guangyey/163/head 2025-12-04T09:20:54.2921949Z * [new branch] gh/guangyey/163/orig -> origin/gh/guangyey/163/orig 2025-12-04T09:20:54.2922512Z * [new branch] gh/guangyey/168/base -> origin/gh/guangyey/168/base 2025-12-04T09:20:54.2922940Z * [new branch] gh/guangyey/168/head -> origin/gh/guangyey/168/head 2025-12-04T09:20:54.2923205Z * [new branch] gh/guangyey/168/orig -> origin/gh/guangyey/168/orig 2025-12-04T09:20:54.2927928Z * [new branch] gh/guangyey/169/base -> origin/gh/guangyey/169/base 2025-12-04T09:20:54.2928134Z * [new branch] gh/guangyey/169/head -> origin/gh/guangyey/169/head 2025-12-04T09:20:54.2928284Z * [new branch] gh/guangyey/169/orig -> origin/gh/guangyey/169/orig 2025-12-04T09:20:54.2928468Z * [new branch] gh/guangyey/170/base -> origin/gh/guangyey/170/base 2025-12-04T09:20:54.2928611Z * [new branch] gh/guangyey/170/head -> origin/gh/guangyey/170/head 2025-12-04T09:20:54.2928753Z * [new branch] gh/guangyey/170/orig -> origin/gh/guangyey/170/orig 2025-12-04T09:20:54.2931769Z * [new branch] gh/guangyey/171/base -> origin/gh/guangyey/171/base 2025-12-04T09:20:54.2932092Z * [new branch] gh/guangyey/171/head -> origin/gh/guangyey/171/head 2025-12-04T09:20:54.2932272Z * [new branch] gh/guangyey/171/orig -> origin/gh/guangyey/171/orig 2025-12-04T09:20:54.2932518Z * [new branch] gh/guangyey/178/base -> origin/gh/guangyey/178/base 2025-12-04T09:20:54.2932677Z * [new branch] gh/guangyey/178/head -> origin/gh/guangyey/178/head 2025-12-04T09:20:54.2933046Z * [new branch] gh/guangyey/178/orig -> origin/gh/guangyey/178/orig 2025-12-04T09:20:54.2937262Z * [new branch] gh/guangyey/182/base -> origin/gh/guangyey/182/base 2025-12-04T09:20:54.2937620Z * [new branch] gh/guangyey/182/head -> origin/gh/guangyey/182/head 2025-12-04T09:20:54.2937867Z * [new branch] gh/guangyey/182/orig -> origin/gh/guangyey/182/orig 2025-12-04T09:20:54.2938080Z * [new branch] gh/guangyey/183/base -> origin/gh/guangyey/183/base 2025-12-04T09:20:54.2938332Z * [new branch] gh/guangyey/183/head -> origin/gh/guangyey/183/head 2025-12-04T09:20:54.2938494Z * [new branch] gh/guangyey/183/orig -> origin/gh/guangyey/183/orig 2025-12-04T09:20:54.2938947Z * [new branch] gh/guangyey/185/base -> origin/gh/guangyey/185/base 2025-12-04T09:20:54.2939467Z * [new branch] gh/guangyey/185/head -> origin/gh/guangyey/185/head 2025-12-04T09:20:54.2945303Z * [new branch] gh/guangyey/185/orig -> origin/gh/guangyey/185/orig 2025-12-04T09:20:54.2950937Z * [new branch] gh/guangyey/186/base -> origin/gh/guangyey/186/base 2025-12-04T09:20:54.2956039Z * [new branch] gh/guangyey/186/head -> origin/gh/guangyey/186/head 2025-12-04T09:20:54.2958368Z * [new branch] gh/guangyey/186/orig -> origin/gh/guangyey/186/orig 2025-12-04T09:20:54.2958964Z * [new branch] gh/guangyey/187/base -> origin/gh/guangyey/187/base 2025-12-04T09:20:54.2959487Z * [new branch] gh/guangyey/187/head -> origin/gh/guangyey/187/head 2025-12-04T09:20:54.2959695Z * [new branch] gh/guangyey/187/orig -> origin/gh/guangyey/187/orig 2025-12-04T09:20:54.2959851Z * [new branch] gh/guangyey/188/base -> origin/gh/guangyey/188/base 2025-12-04T09:20:54.2960000Z * [new branch] gh/guangyey/188/head -> origin/gh/guangyey/188/head 2025-12-04T09:20:54.2960177Z * [new branch] gh/guangyey/188/orig -> origin/gh/guangyey/188/orig 2025-12-04T09:20:54.2960324Z * [new branch] gh/guangyey/190/base -> origin/gh/guangyey/190/base 2025-12-04T09:20:54.2960480Z * [new branch] gh/guangyey/190/head -> origin/gh/guangyey/190/head 2025-12-04T09:20:54.2960654Z * [new branch] gh/guangyey/190/orig -> origin/gh/guangyey/190/orig 2025-12-04T09:20:54.2960800Z * [new branch] gh/guangyey/208/base -> origin/gh/guangyey/208/base 2025-12-04T09:20:54.2960952Z * [new branch] gh/guangyey/208/head -> origin/gh/guangyey/208/head 2025-12-04T09:20:54.2961095Z * [new branch] gh/guangyey/208/orig -> origin/gh/guangyey/208/orig 2025-12-04T09:20:54.2961250Z * [new branch] gh/guangyey/228/base -> origin/gh/guangyey/228/base 2025-12-04T09:20:54.2961396Z * [new branch] gh/guangyey/228/head -> origin/gh/guangyey/228/head 2025-12-04T09:20:54.2961544Z * [new branch] gh/guangyey/228/orig -> origin/gh/guangyey/228/orig 2025-12-04T09:20:54.2961776Z * [new branch] gh/guangyey/230/base -> origin/gh/guangyey/230/base 2025-12-04T09:20:54.2961928Z * [new branch] gh/guangyey/230/head -> origin/gh/guangyey/230/head 2025-12-04T09:20:54.2962075Z * [new branch] gh/guangyey/230/orig -> origin/gh/guangyey/230/orig 2025-12-04T09:20:54.2962228Z * [new branch] gh/guangyey/231/base -> origin/gh/guangyey/231/base 2025-12-04T09:20:54.2962369Z * [new branch] gh/guangyey/231/head -> origin/gh/guangyey/231/head 2025-12-04T09:20:54.2962519Z * [new branch] gh/guangyey/231/orig -> origin/gh/guangyey/231/orig 2025-12-04T09:20:54.2962822Z * [new branch] gh/guangyey/232/base -> origin/gh/guangyey/232/base 2025-12-04T09:20:54.2963138Z * [new branch] gh/guangyey/232/head -> origin/gh/guangyey/232/head 2025-12-04T09:20:54.2963761Z * [new branch] gh/guangyey/232/orig -> origin/gh/guangyey/232/orig 2025-12-04T09:20:54.2963973Z * [new branch] gh/guangyey/233/base -> origin/gh/guangyey/233/base 2025-12-04T09:20:54.2964148Z * [new branch] gh/guangyey/233/head -> origin/gh/guangyey/233/head 2025-12-04T09:20:54.2964321Z * [new branch] gh/guangyey/233/orig -> origin/gh/guangyey/233/orig 2025-12-04T09:20:54.2969725Z * [new branch] gh/guangyey/234/base -> origin/gh/guangyey/234/base 2025-12-04T09:20:54.2974391Z * [new branch] gh/guangyey/234/head -> origin/gh/guangyey/234/head 2025-12-04T09:20:54.2976490Z * [new branch] gh/guangyey/234/orig -> origin/gh/guangyey/234/orig 2025-12-04T09:20:54.2976704Z * [new branch] gh/guangyey/235/base -> origin/gh/guangyey/235/base 2025-12-04T09:20:54.2977199Z * [new branch] gh/guangyey/235/head -> origin/gh/guangyey/235/head 2025-12-04T09:20:54.2977384Z * [new branch] gh/guangyey/235/orig -> origin/gh/guangyey/235/orig 2025-12-04T09:20:54.2978109Z * [new branch] gh/guangyey/236/base -> origin/gh/guangyey/236/base 2025-12-04T09:20:54.2978453Z * [new branch] gh/guangyey/236/head -> origin/gh/guangyey/236/head 2025-12-04T09:20:54.2978625Z * [new branch] gh/guangyey/236/orig -> origin/gh/guangyey/236/orig 2025-12-04T09:20:54.2978783Z * [new branch] gh/guangyey/237/base -> origin/gh/guangyey/237/base 2025-12-04T09:20:54.2978932Z * [new branch] gh/guangyey/237/head -> origin/gh/guangyey/237/head 2025-12-04T09:20:54.2979080Z * [new branch] gh/guangyey/237/orig -> origin/gh/guangyey/237/orig 2025-12-04T09:20:54.2979231Z * [new branch] gh/guangyey/238/base -> origin/gh/guangyey/238/base 2025-12-04T09:20:54.2979380Z * [new branch] gh/guangyey/238/head -> origin/gh/guangyey/238/head 2025-12-04T09:20:54.2979536Z * [new branch] gh/guangyey/239/base -> origin/gh/guangyey/239/base 2025-12-04T09:20:54.2979679Z * [new branch] gh/guangyey/239/head -> origin/gh/guangyey/239/head 2025-12-04T09:20:54.2979824Z * [new branch] gh/guangyey/239/orig -> origin/gh/guangyey/239/orig 2025-12-04T09:20:54.2980145Z * [new branch] gh/guangyey/240/base -> origin/gh/guangyey/240/base 2025-12-04T09:20:54.2980307Z * [new branch] gh/guangyey/240/head -> origin/gh/guangyey/240/head 2025-12-04T09:20:54.2981166Z * [new branch] gh/guangyey/240/orig -> origin/gh/guangyey/240/orig 2025-12-04T09:20:54.2984654Z * [new branch] gh/guangyey/241/base -> origin/gh/guangyey/241/base 2025-12-04T09:20:54.2985062Z * [new branch] gh/guangyey/241/head -> origin/gh/guangyey/241/head 2025-12-04T09:20:54.2985292Z * [new branch] gh/guangyey/241/orig -> origin/gh/guangyey/241/orig 2025-12-04T09:20:54.2985589Z * [new branch] gh/guangyey/242/base -> origin/gh/guangyey/242/base 2025-12-04T09:20:54.2985819Z * [new branch] gh/guangyey/242/head -> origin/gh/guangyey/242/head 2025-12-04T09:20:54.2986268Z * [new branch] gh/guangyey/242/orig -> origin/gh/guangyey/242/orig 2025-12-04T09:20:54.2990062Z * [new branch] gh/guangyey/243/base -> origin/gh/guangyey/243/base 2025-12-04T09:20:54.2990479Z * [new branch] gh/guangyey/243/head -> origin/gh/guangyey/243/head 2025-12-04T09:20:54.2990863Z * [new branch] gh/guangyey/243/orig -> origin/gh/guangyey/243/orig 2025-12-04T09:20:54.2991020Z * [new branch] gh/guangyey/244/base -> origin/gh/guangyey/244/base 2025-12-04T09:20:54.2991318Z * [new branch] gh/guangyey/244/head -> origin/gh/guangyey/244/head 2025-12-04T09:20:54.2991564Z * [new branch] gh/guangyey/244/orig -> origin/gh/guangyey/244/orig 2025-12-04T09:20:54.2993029Z * [new branch] gh/guangyey/245/base -> origin/gh/guangyey/245/base 2025-12-04T09:20:54.2993202Z * [new branch] gh/guangyey/245/head -> origin/gh/guangyey/245/head 2025-12-04T09:20:54.2995772Z * [new branch] gh/guangyey/245/orig -> origin/gh/guangyey/245/orig 2025-12-04T09:20:54.2996130Z * [new branch] gh/guangyey/246/base -> origin/gh/guangyey/246/base 2025-12-04T09:20:54.2996317Z * [new branch] gh/guangyey/246/head -> origin/gh/guangyey/246/head 2025-12-04T09:20:54.2996510Z * [new branch] gh/guangyey/246/orig -> origin/gh/guangyey/246/orig 2025-12-04T09:20:54.2998326Z * [new branch] gh/guangyey/247/base -> origin/gh/guangyey/247/base 2025-12-04T09:20:54.2998636Z * [new branch] gh/guangyey/247/head -> origin/gh/guangyey/247/head 2025-12-04T09:20:54.2998797Z * [new branch] gh/guangyey/247/orig -> origin/gh/guangyey/247/orig 2025-12-04T09:20:54.3000552Z * [new branch] gh/guangyey/248/base -> origin/gh/guangyey/248/base 2025-12-04T09:20:54.3001221Z * [new branch] gh/guangyey/248/head -> origin/gh/guangyey/248/head 2025-12-04T09:20:54.3001580Z * [new branch] gh/guangyey/248/orig -> origin/gh/guangyey/248/orig 2025-12-04T09:20:54.3002160Z * [new branch] gh/guangyey/249/base -> origin/gh/guangyey/249/base 2025-12-04T09:20:54.3006138Z * [new branch] gh/guangyey/249/head -> origin/gh/guangyey/249/head 2025-12-04T09:20:54.3006531Z * [new branch] gh/guangyey/249/orig -> origin/gh/guangyey/249/orig 2025-12-04T09:20:54.3006707Z * [new branch] gh/guangyey/250/base -> origin/gh/guangyey/250/base 2025-12-04T09:20:54.3006882Z * [new branch] gh/guangyey/250/head -> origin/gh/guangyey/250/head 2025-12-04T09:20:54.3007172Z * [new branch] gh/guangyey/250/orig -> origin/gh/guangyey/250/orig 2025-12-04T09:20:54.3007334Z * [new branch] gh/guangyey/251/base -> origin/gh/guangyey/251/base 2025-12-04T09:20:54.3007511Z * [new branch] gh/guangyey/251/head -> origin/gh/guangyey/251/head 2025-12-04T09:20:54.3008497Z * [new branch] gh/guangyey/251/orig -> origin/gh/guangyey/251/orig 2025-12-04T09:20:54.3011960Z * [new branch] gh/guangyey/252/base -> origin/gh/guangyey/252/base 2025-12-04T09:20:54.3012307Z * [new branch] gh/guangyey/252/head -> origin/gh/guangyey/252/head 2025-12-04T09:20:54.3012503Z * [new branch] gh/guangyey/252/orig -> origin/gh/guangyey/252/orig 2025-12-04T09:20:54.3012659Z * [new branch] gh/guangyey/253/base -> origin/gh/guangyey/253/base 2025-12-04T09:20:54.3018708Z * [new branch] gh/guangyey/253/head -> origin/gh/guangyey/253/head 2025-12-04T09:20:54.3018888Z * [new branch] gh/guangyey/253/orig -> origin/gh/guangyey/253/orig 2025-12-04T09:20:54.3019098Z * [new branch] gh/guangyey/254/base -> origin/gh/guangyey/254/base 2025-12-04T09:20:54.3019279Z * [new branch] gh/guangyey/254/head -> origin/gh/guangyey/254/head 2025-12-04T09:20:54.3019426Z * [new branch] gh/guangyey/254/orig -> origin/gh/guangyey/254/orig 2025-12-04T09:20:54.3019580Z * [new branch] gh/guangyey/255/base -> origin/gh/guangyey/255/base 2025-12-04T09:20:54.3019724Z * [new branch] gh/guangyey/255/head -> origin/gh/guangyey/255/head 2025-12-04T09:20:54.3019875Z * [new branch] gh/guangyey/255/orig -> origin/gh/guangyey/255/orig 2025-12-04T09:20:54.3021997Z * [new branch] gh/guilhermeleobas/107/base -> origin/gh/guilhermeleobas/107/base 2025-12-04T09:20:54.3022443Z * [new branch] gh/guilhermeleobas/107/head -> origin/gh/guilhermeleobas/107/head 2025-12-04T09:20:54.3022643Z * [new branch] gh/guilhermeleobas/107/orig -> origin/gh/guilhermeleobas/107/orig 2025-12-04T09:20:54.3022846Z * [new branch] gh/guilhermeleobas/108/base -> origin/gh/guilhermeleobas/108/base 2025-12-04T09:20:54.3023020Z * [new branch] gh/guilhermeleobas/108/head -> origin/gh/guilhermeleobas/108/head 2025-12-04T09:20:54.3023199Z * [new branch] gh/guilhermeleobas/108/orig -> origin/gh/guilhermeleobas/108/orig 2025-12-04T09:20:54.3028405Z * [new branch] gh/guilhermeleobas/150/base -> origin/gh/guilhermeleobas/150/base 2025-12-04T09:20:54.3028594Z * [new branch] gh/guilhermeleobas/150/head -> origin/gh/guilhermeleobas/150/head 2025-12-04T09:20:54.3028775Z * [new branch] gh/guilhermeleobas/150/orig -> origin/gh/guilhermeleobas/150/orig 2025-12-04T09:20:54.3029084Z * [new branch] gh/guilhermeleobas/168/base -> origin/gh/guilhermeleobas/168/base 2025-12-04T09:20:54.3029261Z * [new branch] gh/guilhermeleobas/168/head -> origin/gh/guilhermeleobas/168/head 2025-12-04T09:20:54.3029430Z * [new branch] gh/guilhermeleobas/168/orig -> origin/gh/guilhermeleobas/168/orig 2025-12-04T09:20:54.3033701Z * [new branch] gh/guilhermeleobas/169/base -> origin/gh/guilhermeleobas/169/base 2025-12-04T09:20:54.3033957Z * [new branch] gh/guilhermeleobas/169/head -> origin/gh/guilhermeleobas/169/head 2025-12-04T09:20:54.3034170Z * [new branch] gh/guilhermeleobas/169/orig -> origin/gh/guilhermeleobas/169/orig 2025-12-04T09:20:54.3034353Z * [new branch] gh/guilhermeleobas/170/base -> origin/gh/guilhermeleobas/170/base 2025-12-04T09:20:54.3034521Z * [new branch] gh/guilhermeleobas/170/head -> origin/gh/guilhermeleobas/170/head 2025-12-04T09:20:54.3037113Z * [new branch] gh/guilhermeleobas/170/orig -> origin/gh/guilhermeleobas/170/orig 2025-12-04T09:20:54.3037333Z * [new branch] gh/guilhermeleobas/171/base -> origin/gh/guilhermeleobas/171/base 2025-12-04T09:20:54.3038043Z * [new branch] gh/guilhermeleobas/171/head -> origin/gh/guilhermeleobas/171/head 2025-12-04T09:20:54.3038271Z * [new branch] gh/guilhermeleobas/171/orig -> origin/gh/guilhermeleobas/171/orig 2025-12-04T09:20:54.3038515Z * [new branch] gh/guilhermeleobas/173/base -> origin/gh/guilhermeleobas/173/base 2025-12-04T09:20:54.3038702Z * [new branch] gh/guilhermeleobas/173/head -> origin/gh/guilhermeleobas/173/head 2025-12-04T09:20:54.3038917Z * [new branch] gh/guilhermeleobas/173/orig -> origin/gh/guilhermeleobas/173/orig 2025-12-04T09:20:54.3039193Z * [new branch] gh/guilhermeleobas/193/base -> origin/gh/guilhermeleobas/193/base 2025-12-04T09:20:54.3039749Z * [new branch] gh/guilhermeleobas/193/head -> origin/gh/guilhermeleobas/193/head 2025-12-04T09:20:54.3040779Z * [new branch] gh/guilhermeleobas/193/orig -> origin/gh/guilhermeleobas/193/orig 2025-12-04T09:20:54.3041882Z * [new branch] gh/guilhermeleobas/204/base -> origin/gh/guilhermeleobas/204/base 2025-12-04T09:20:54.3043154Z * [new branch] gh/guilhermeleobas/204/head -> origin/gh/guilhermeleobas/204/head 2025-12-04T09:20:54.3043410Z * [new branch] gh/guilhermeleobas/204/orig -> origin/gh/guilhermeleobas/204/orig 2025-12-04T09:20:54.3048777Z * [new branch] gh/guilhermeleobas/211/base -> origin/gh/guilhermeleobas/211/base 2025-12-04T09:20:54.3048993Z * [new branch] gh/guilhermeleobas/211/head -> origin/gh/guilhermeleobas/211/head 2025-12-04T09:20:54.3049188Z * [new branch] gh/guilhermeleobas/211/orig -> origin/gh/guilhermeleobas/211/orig 2025-12-04T09:20:54.3049368Z * [new branch] gh/guilhermeleobas/226/base -> origin/gh/guilhermeleobas/226/base 2025-12-04T09:20:54.3049570Z * [new branch] gh/guilhermeleobas/226/head -> origin/gh/guilhermeleobas/226/head 2025-12-04T09:20:54.3049751Z * [new branch] gh/guilhermeleobas/226/orig -> origin/gh/guilhermeleobas/226/orig 2025-12-04T09:20:54.3049958Z * [new branch] gh/guilhermeleobas/236/base -> origin/gh/guilhermeleobas/236/base 2025-12-04T09:20:54.3050382Z * [new branch] gh/guilhermeleobas/236/head -> origin/gh/guilhermeleobas/236/head 2025-12-04T09:20:54.3050771Z * [new branch] gh/guilhermeleobas/236/orig -> origin/gh/guilhermeleobas/236/orig 2025-12-04T09:20:54.3055011Z * [new branch] gh/guilhermeleobas/247/base -> origin/gh/guilhermeleobas/247/base 2025-12-04T09:20:54.3055228Z * [new branch] gh/guilhermeleobas/247/head -> origin/gh/guilhermeleobas/247/head 2025-12-04T09:20:54.3055407Z * [new branch] gh/guilhermeleobas/247/orig -> origin/gh/guilhermeleobas/247/orig 2025-12-04T09:20:54.3055840Z * [new branch] gh/guilhermeleobas/248/base -> origin/gh/guilhermeleobas/248/base 2025-12-04T09:20:54.3056019Z * [new branch] gh/guilhermeleobas/248/head -> origin/gh/guilhermeleobas/248/head 2025-12-04T09:20:54.3056197Z * [new branch] gh/guilhermeleobas/248/orig -> origin/gh/guilhermeleobas/248/orig 2025-12-04T09:20:54.3056964Z * [new branch] gh/guilhermeleobas/250/base -> origin/gh/guilhermeleobas/250/base 2025-12-04T09:20:54.3057582Z * [new branch] gh/guilhermeleobas/250/head -> origin/gh/guilhermeleobas/250/head 2025-12-04T09:20:54.3058542Z * [new branch] gh/guilhermeleobas/250/orig -> origin/gh/guilhermeleobas/250/orig 2025-12-04T09:20:54.3062090Z * [new branch] gh/guilhermeleobas/253/base -> origin/gh/guilhermeleobas/253/base 2025-12-04T09:20:54.3062653Z * [new branch] gh/guilhermeleobas/253/head -> origin/gh/guilhermeleobas/253/head 2025-12-04T09:20:54.3062879Z * [new branch] gh/guilhermeleobas/253/orig -> origin/gh/guilhermeleobas/253/orig 2025-12-04T09:20:54.3063085Z * [new branch] gh/guilhermeleobas/254/base -> origin/gh/guilhermeleobas/254/base 2025-12-04T09:20:54.3064170Z * [new branch] gh/guilhermeleobas/254/head -> origin/gh/guilhermeleobas/254/head 2025-12-04T09:20:54.3064376Z * [new branch] gh/guilhermeleobas/254/orig -> origin/gh/guilhermeleobas/254/orig 2025-12-04T09:20:54.3064674Z * [new branch] gh/guilhermeleobas/255/base -> origin/gh/guilhermeleobas/255/base 2025-12-04T09:20:54.3066976Z * [new branch] gh/guilhermeleobas/255/head -> origin/gh/guilhermeleobas/255/head 2025-12-04T09:20:54.3067202Z * [new branch] gh/guilhermeleobas/255/orig -> origin/gh/guilhermeleobas/255/orig 2025-12-04T09:20:54.3067377Z * [new branch] gh/guilhermeleobas/256/base -> origin/gh/guilhermeleobas/256/base 2025-12-04T09:20:54.3068067Z * [new branch] gh/guilhermeleobas/256/head -> origin/gh/guilhermeleobas/256/head 2025-12-04T09:20:54.3068811Z * [new branch] gh/guilhermeleobas/256/orig -> origin/gh/guilhermeleobas/256/orig 2025-12-04T09:20:54.3072097Z * [new branch] gh/guilhermeleobas/257/base -> origin/gh/guilhermeleobas/257/base 2025-12-04T09:20:54.3072331Z * [new branch] gh/guilhermeleobas/257/head -> origin/gh/guilhermeleobas/257/head 2025-12-04T09:20:54.3072506Z * [new branch] gh/guilhermeleobas/257/orig -> origin/gh/guilhermeleobas/257/orig 2025-12-04T09:20:54.3072680Z * [new branch] gh/guilhermeleobas/258/base -> origin/gh/guilhermeleobas/258/base 2025-12-04T09:20:54.3073094Z * [new branch] gh/guilhermeleobas/258/head -> origin/gh/guilhermeleobas/258/head 2025-12-04T09:20:54.3073940Z * [new branch] gh/guilhermeleobas/258/orig -> origin/gh/guilhermeleobas/258/orig 2025-12-04T09:20:54.3075004Z * [new branch] gh/guilhermeleobas/259/base -> origin/gh/guilhermeleobas/259/base 2025-12-04T09:20:54.3075592Z * [new branch] gh/guilhermeleobas/259/head -> origin/gh/guilhermeleobas/259/head 2025-12-04T09:20:54.3076433Z * [new branch] gh/guilhermeleobas/259/orig -> origin/gh/guilhermeleobas/259/orig 2025-12-04T09:20:54.3077470Z * [new branch] gh/guilhermeleobas/260/base -> origin/gh/guilhermeleobas/260/base 2025-12-04T09:20:54.3077973Z * [new branch] gh/guilhermeleobas/260/head -> origin/gh/guilhermeleobas/260/head 2025-12-04T09:20:54.3078918Z * [new branch] gh/guilhermeleobas/260/orig -> origin/gh/guilhermeleobas/260/orig 2025-12-04T09:20:54.3079944Z * [new branch] gh/guilhermeleobas/261/base -> origin/gh/guilhermeleobas/261/base 2025-12-04T09:20:54.3080643Z * [new branch] gh/guilhermeleobas/261/head -> origin/gh/guilhermeleobas/261/head 2025-12-04T09:20:54.3081337Z * [new branch] gh/guilhermeleobas/261/orig -> origin/gh/guilhermeleobas/261/orig 2025-12-04T09:20:54.3082319Z * [new branch] gh/guilhermeleobas/262/base -> origin/gh/guilhermeleobas/262/base 2025-12-04T09:20:54.3084643Z * [new branch] gh/guilhermeleobas/262/head -> origin/gh/guilhermeleobas/262/head 2025-12-04T09:20:54.3084858Z * [new branch] gh/guilhermeleobas/262/orig -> origin/gh/guilhermeleobas/262/orig 2025-12-04T09:20:54.3085192Z * [new branch] gh/guilhermeleobas/263/base -> origin/gh/guilhermeleobas/263/base 2025-12-04T09:20:54.3085608Z * [new branch] gh/guilhermeleobas/263/head -> origin/gh/guilhermeleobas/263/head 2025-12-04T09:20:54.3086329Z * [new branch] gh/guilhermeleobas/263/orig -> origin/gh/guilhermeleobas/263/orig 2025-12-04T09:20:54.3090183Z * [new branch] gh/guilhermeleobas/264/base -> origin/gh/guilhermeleobas/264/base 2025-12-04T09:20:54.3091948Z * [new branch] gh/guilhermeleobas/264/head -> origin/gh/guilhermeleobas/264/head 2025-12-04T09:20:54.3092576Z * [new branch] gh/guilhermeleobas/264/orig -> origin/gh/guilhermeleobas/264/orig 2025-12-04T09:20:54.3092806Z * [new branch] gh/guilhermeleobas/265/base -> origin/gh/guilhermeleobas/265/base 2025-12-04T09:20:54.3093201Z * [new branch] gh/guilhermeleobas/265/head -> origin/gh/guilhermeleobas/265/head 2025-12-04T09:20:54.3093394Z * [new branch] gh/guilhermeleobas/265/orig -> origin/gh/guilhermeleobas/265/orig 2025-12-04T09:20:54.3093572Z * [new branch] gh/guilhermeleobas/266/base -> origin/gh/guilhermeleobas/266/base 2025-12-04T09:20:54.3093742Z * [new branch] gh/guilhermeleobas/266/head -> origin/gh/guilhermeleobas/266/head 2025-12-04T09:20:54.3098212Z * [new branch] gh/guilhermeleobas/266/orig -> origin/gh/guilhermeleobas/266/orig 2025-12-04T09:20:54.3098393Z * [new branch] gh/guilhermeleobas/267/base -> origin/gh/guilhermeleobas/267/base 2025-12-04T09:20:54.3098565Z * [new branch] gh/guilhermeleobas/267/head -> origin/gh/guilhermeleobas/267/head 2025-12-04T09:20:54.3098735Z * [new branch] gh/guilhermeleobas/267/orig -> origin/gh/guilhermeleobas/267/orig 2025-12-04T09:20:54.3098906Z * [new branch] gh/hameerabbasi/1/base -> origin/gh/hameerabbasi/1/base 2025-12-04T09:20:54.3099068Z * [new branch] gh/hameerabbasi/1/head -> origin/gh/hameerabbasi/1/head 2025-12-04T09:20:54.3101273Z * [new branch] gh/hameerabbasi/2/base -> origin/gh/hameerabbasi/2/base 2025-12-04T09:20:54.3101987Z * [new branch] gh/hameerabbasi/2/head -> origin/gh/hameerabbasi/2/head 2025-12-04T09:20:54.3102260Z * [new branch] gh/hameerabbasi/2/orig -> origin/gh/hameerabbasi/2/orig 2025-12-04T09:20:54.3102435Z * [new branch] gh/hameerabbasi/3/base -> origin/gh/hameerabbasi/3/base 2025-12-04T09:20:54.3105652Z * [new branch] gh/hameerabbasi/3/head -> origin/gh/hameerabbasi/3/head 2025-12-04T09:20:54.3105925Z * [new branch] gh/hameerabbasi/3/orig -> origin/gh/hameerabbasi/3/orig 2025-12-04T09:20:54.3106099Z * [new branch] gh/hameerabbasi/4/base -> origin/gh/hameerabbasi/4/base 2025-12-04T09:20:54.3106260Z * [new branch] gh/hameerabbasi/4/head -> origin/gh/hameerabbasi/4/head 2025-12-04T09:20:54.3106452Z * [new branch] gh/hameerabbasi/4/orig -> origin/gh/hameerabbasi/4/orig 2025-12-04T09:20:54.3109714Z * [new branch] gh/huydhn/1/next -> origin/gh/huydhn/1/next 2025-12-04T09:20:54.3109872Z * [new branch] gh/huydhn/2/next -> origin/gh/huydhn/2/next 2025-12-04T09:20:54.3110016Z * [new branch] gh/huydhn/3/next -> origin/gh/huydhn/3/next 2025-12-04T09:20:54.3114051Z * [new branch] gh/huydhn/4/next -> origin/gh/huydhn/4/next 2025-12-04T09:20:54.3114305Z * [new branch] gh/huydhn/5/next -> origin/gh/huydhn/5/next 2025-12-04T09:20:54.3114603Z * [new branch] gh/huydhn/6/next -> origin/gh/huydhn/6/next 2025-12-04T09:20:54.3114765Z * [new branch] gh/int3/97/base -> origin/gh/int3/97/base 2025-12-04T09:20:54.3117760Z * [new branch] gh/int3/97/head -> origin/gh/int3/97/head 2025-12-04T09:20:54.3118002Z * [new branch] gh/isuruf/101/base -> origin/gh/isuruf/101/base 2025-12-04T09:20:54.3118159Z * [new branch] gh/isuruf/101/head -> origin/gh/isuruf/101/head 2025-12-04T09:20:54.3118305Z * [new branch] gh/isuruf/146/base -> origin/gh/isuruf/146/base 2025-12-04T09:20:54.3118777Z * [new branch] gh/isuruf/146/head -> origin/gh/isuruf/146/head 2025-12-04T09:20:54.3119214Z * [new branch] gh/isuruf/146/orig -> origin/gh/isuruf/146/orig 2025-12-04T09:20:54.3120543Z * [new branch] gh/isuruf/158/base -> origin/gh/isuruf/158/base 2025-12-04T09:20:54.3120889Z * [new branch] gh/isuruf/158/head -> origin/gh/isuruf/158/head 2025-12-04T09:20:54.3126103Z * [new branch] gh/isuruf/159/base -> origin/gh/isuruf/159/base 2025-12-04T09:20:54.3126647Z * [new branch] gh/isuruf/159/head -> origin/gh/isuruf/159/head 2025-12-04T09:20:54.3126855Z * [new branch] gh/isuruf/160/base -> origin/gh/isuruf/160/base 2025-12-04T09:20:54.3127015Z * [new branch] gh/isuruf/160/head -> origin/gh/isuruf/160/head 2025-12-04T09:20:54.3127163Z * [new branch] gh/isuruf/160/orig -> origin/gh/isuruf/160/orig 2025-12-04T09:20:54.3127319Z * [new branch] gh/isuruf/81/base -> origin/gh/isuruf/81/base 2025-12-04T09:20:54.3127464Z * [new branch] gh/isuruf/81/head -> origin/gh/isuruf/81/head 2025-12-04T09:20:54.3127823Z * [new branch] gh/isuruf/81/orig -> origin/gh/isuruf/81/orig 2025-12-04T09:20:54.3129227Z * [new branch] gh/jamesjwu/176/base -> origin/gh/jamesjwu/176/base 2025-12-04T09:20:54.3129617Z * [new branch] gh/jamesjwu/176/head -> origin/gh/jamesjwu/176/head 2025-12-04T09:20:54.3133135Z * [new branch] gh/jamesjwu/176/orig -> origin/gh/jamesjwu/176/orig 2025-12-04T09:20:54.3134854Z * [new branch] gh/jamesjwu/187/base -> origin/gh/jamesjwu/187/base 2025-12-04T09:20:54.3135201Z * [new branch] gh/jamesjwu/187/head -> origin/gh/jamesjwu/187/head 2025-12-04T09:20:54.3138041Z * [new branch] gh/jamesjwu/187/orig -> origin/gh/jamesjwu/187/orig 2025-12-04T09:20:54.3138269Z * [new branch] gh/jamesjwu/196/base -> origin/gh/jamesjwu/196/base 2025-12-04T09:20:54.3138538Z * [new branch] gh/jamesjwu/196/head -> origin/gh/jamesjwu/196/head 2025-12-04T09:20:54.3138770Z * [new branch] gh/jamesjwu/196/orig -> origin/gh/jamesjwu/196/orig 2025-12-04T09:20:54.3142971Z * [new branch] gh/jamesjwu/198/base -> origin/gh/jamesjwu/198/base 2025-12-04T09:20:54.3143172Z * [new branch] gh/jamesjwu/198/head -> origin/gh/jamesjwu/198/head 2025-12-04T09:20:54.3143328Z * [new branch] gh/jamesjwu/198/orig -> origin/gh/jamesjwu/198/orig 2025-12-04T09:20:54.3143492Z * [new branch] gh/jamesjwu/207/base -> origin/gh/jamesjwu/207/base 2025-12-04T09:20:54.3143652Z * [new branch] gh/jamesjwu/207/head -> origin/gh/jamesjwu/207/head 2025-12-04T09:20:54.3144112Z * [new branch] gh/jamesjwu/207/orig -> origin/gh/jamesjwu/207/orig 2025-12-04T09:20:54.3148098Z * [new branch] gh/jamesjwu/208/base -> origin/gh/jamesjwu/208/base 2025-12-04T09:20:54.3148287Z * [new branch] gh/jamesjwu/208/head -> origin/gh/jamesjwu/208/head 2025-12-04T09:20:54.3148656Z * [new branch] gh/jamesjwu/208/orig -> origin/gh/jamesjwu/208/orig 2025-12-04T09:20:54.3148829Z * [new branch] gh/jamesjwu/52/base -> origin/gh/jamesjwu/52/base 2025-12-04T09:20:54.3148977Z * [new branch] gh/jamesjwu/52/head -> origin/gh/jamesjwu/52/head 2025-12-04T09:20:54.3149462Z * [new branch] gh/jamesjwu/53/base -> origin/gh/jamesjwu/53/base 2025-12-04T09:20:54.3152497Z * [new branch] gh/jamesjwu/53/head -> origin/gh/jamesjwu/53/head 2025-12-04T09:20:54.3152838Z * [new branch] gh/jamesjwu/54/base -> origin/gh/jamesjwu/54/base 2025-12-04T09:20:54.3153021Z * [new branch] gh/jamesjwu/54/head -> origin/gh/jamesjwu/54/head 2025-12-04T09:20:54.3153268Z * [new branch] gh/jamesjwu/55/base -> origin/gh/jamesjwu/55/base 2025-12-04T09:20:54.3153580Z * [new branch] gh/jamesjwu/55/head -> origin/gh/jamesjwu/55/head 2025-12-04T09:20:54.3158770Z * [new branch] gh/jamesjwu/56/base -> origin/gh/jamesjwu/56/base 2025-12-04T09:20:54.3159313Z * [new branch] gh/jamesjwu/56/head -> origin/gh/jamesjwu/56/head 2025-12-04T09:20:54.3159500Z * [new branch] gh/jamesjwu/57/base -> origin/gh/jamesjwu/57/base 2025-12-04T09:20:54.3159674Z * [new branch] gh/jamesjwu/57/head -> origin/gh/jamesjwu/57/head 2025-12-04T09:20:54.3159822Z * [new branch] gh/jamesjwu/58/base -> origin/gh/jamesjwu/58/base 2025-12-04T09:20:54.3159968Z * [new branch] gh/jamesjwu/58/head -> origin/gh/jamesjwu/58/head 2025-12-04T09:20:54.3160121Z * [new branch] gh/jamesjwu/59/base -> origin/gh/jamesjwu/59/base 2025-12-04T09:20:54.3160300Z * [new branch] gh/jamesjwu/59/head -> origin/gh/jamesjwu/59/head 2025-12-04T09:20:54.3162069Z * [new branch] gh/jamesjwu/60/base -> origin/gh/jamesjwu/60/base 2025-12-04T09:20:54.3162286Z * [new branch] gh/jamesjwu/60/head -> origin/gh/jamesjwu/60/head 2025-12-04T09:20:54.3168172Z * [new branch] gh/jamesjwu/61/base -> origin/gh/jamesjwu/61/base 2025-12-04T09:20:54.3168724Z * [new branch] gh/jamesjwu/61/head -> origin/gh/jamesjwu/61/head 2025-12-04T09:20:54.3168912Z * [new branch] gh/jamesjwu/62/base -> origin/gh/jamesjwu/62/base 2025-12-04T09:20:54.3169061Z * [new branch] gh/jamesjwu/62/head -> origin/gh/jamesjwu/62/head 2025-12-04T09:20:54.3169226Z * [new branch] gh/jamesjwu/63/base -> origin/gh/jamesjwu/63/base 2025-12-04T09:20:54.3174328Z * [new branch] gh/jamesjwu/63/head -> origin/gh/jamesjwu/63/head 2025-12-04T09:20:54.3174991Z * [new branch] gh/jamesjwu/64/base -> origin/gh/jamesjwu/64/base 2025-12-04T09:20:54.3175180Z * [new branch] gh/jamesjwu/64/head -> origin/gh/jamesjwu/64/head 2025-12-04T09:20:54.3175348Z * [new branch] gh/jamesjwu/65/base -> origin/gh/jamesjwu/65/base 2025-12-04T09:20:54.3175506Z * [new branch] gh/jamesjwu/65/head -> origin/gh/jamesjwu/65/head 2025-12-04T09:20:54.3175663Z * [new branch] gh/janeyx99/165/base -> origin/gh/janeyx99/165/base 2025-12-04T09:20:54.3175825Z * [new branch] gh/janeyx99/165/head -> origin/gh/janeyx99/165/head 2025-12-04T09:20:54.3175970Z * [new branch] gh/janeyx99/165/orig -> origin/gh/janeyx99/165/orig 2025-12-04T09:20:54.3176112Z * [new branch] gh/janeyx99/201/base -> origin/gh/janeyx99/201/base 2025-12-04T09:20:54.3176262Z * [new branch] gh/janeyx99/201/head -> origin/gh/janeyx99/201/head 2025-12-04T09:20:54.3176408Z * [new branch] gh/janeyx99/201/orig -> origin/gh/janeyx99/201/orig 2025-12-04T09:20:54.3179186Z * [new branch] gh/janeyx99/225/base -> origin/gh/janeyx99/225/base 2025-12-04T09:20:54.3179327Z * [new branch] gh/janeyx99/225/head -> origin/gh/janeyx99/225/head 2025-12-04T09:20:54.3179472Z * [new branch] gh/janeyx99/225/orig -> origin/gh/janeyx99/225/orig 2025-12-04T09:20:54.3179695Z * [new branch] gh/janeyx99/299/base -> origin/gh/janeyx99/299/base 2025-12-04T09:20:54.3179899Z * [new branch] gh/janeyx99/299/head -> origin/gh/janeyx99/299/head 2025-12-04T09:20:54.3186125Z * [new branch] gh/janeyx99/299/orig -> origin/gh/janeyx99/299/orig 2025-12-04T09:20:54.3186322Z * [new branch] gh/janeyx99/302/base -> origin/gh/janeyx99/302/base 2025-12-04T09:20:54.3186645Z * [new branch] gh/janeyx99/302/head -> origin/gh/janeyx99/302/head 2025-12-04T09:20:54.3186792Z * [new branch] gh/janeyx99/303/base -> origin/gh/janeyx99/303/base 2025-12-04T09:20:54.3186951Z * [new branch] gh/janeyx99/303/head -> origin/gh/janeyx99/303/head 2025-12-04T09:20:54.3187092Z * [new branch] gh/janeyx99/305/base -> origin/gh/janeyx99/305/base 2025-12-04T09:20:54.3189072Z * [new branch] gh/janeyx99/305/head -> origin/gh/janeyx99/305/head 2025-12-04T09:20:54.3189341Z * [new branch] gh/janeyx99/306/base -> origin/gh/janeyx99/306/base 2025-12-04T09:20:54.3189504Z * [new branch] gh/janeyx99/306/head -> origin/gh/janeyx99/306/head 2025-12-04T09:20:54.3189728Z * [new branch] gh/janeyx99/314/base -> origin/gh/janeyx99/314/base 2025-12-04T09:20:54.3189880Z * [new branch] gh/janeyx99/314/head -> origin/gh/janeyx99/314/head 2025-12-04T09:20:54.3190066Z * [new branch] gh/janeyx99/314/orig -> origin/gh/janeyx99/314/orig 2025-12-04T09:20:54.3196133Z * [new branch] gh/janeyx99/315/base -> origin/gh/janeyx99/315/base 2025-12-04T09:20:54.3196467Z * [new branch] gh/janeyx99/315/head -> origin/gh/janeyx99/315/head 2025-12-04T09:20:54.3196646Z * [new branch] gh/janeyx99/315/orig -> origin/gh/janeyx99/315/orig 2025-12-04T09:20:54.3196806Z * [new branch] gh/janeyx99/316/base -> origin/gh/janeyx99/316/base 2025-12-04T09:20:54.3196978Z * [new branch] gh/janeyx99/316/head -> origin/gh/janeyx99/316/head 2025-12-04T09:20:54.3202105Z * [new branch] gh/janeyx99/316/orig -> origin/gh/janeyx99/316/orig 2025-12-04T09:20:54.3202324Z * [new branch] gh/janeyx99/317/base -> origin/gh/janeyx99/317/base 2025-12-04T09:20:54.3202475Z * [new branch] gh/janeyx99/317/head -> origin/gh/janeyx99/317/head 2025-12-04T09:20:54.3202626Z * [new branch] gh/janeyx99/317/orig -> origin/gh/janeyx99/317/orig 2025-12-04T09:20:54.3202782Z * [new branch] gh/janeyx99/325/base -> origin/gh/janeyx99/325/base 2025-12-04T09:20:54.3202958Z * [new branch] gh/janeyx99/325/head -> origin/gh/janeyx99/325/head 2025-12-04T09:20:54.3203111Z * [new branch] gh/janeyx99/325/orig -> origin/gh/janeyx99/325/orig 2025-12-04T09:20:54.3203257Z * [new branch] gh/janeyx99/327/base -> origin/gh/janeyx99/327/base 2025-12-04T09:20:54.3203414Z * [new branch] gh/janeyx99/327/head -> origin/gh/janeyx99/327/head 2025-12-04T09:20:54.3203562Z * [new branch] gh/janeyx99/327/orig -> origin/gh/janeyx99/327/orig 2025-12-04T09:20:54.3208670Z * [new branch] gh/janeyx99/328/base -> origin/gh/janeyx99/328/base 2025-12-04T09:20:54.3209122Z * [new branch] gh/janeyx99/328/head -> origin/gh/janeyx99/328/head 2025-12-04T09:20:54.3209303Z * [new branch] gh/janeyx99/328/orig -> origin/gh/janeyx99/328/orig 2025-12-04T09:20:54.3209464Z * [new branch] gh/janeyx99/329/base -> origin/gh/janeyx99/329/base 2025-12-04T09:20:54.3209871Z * [new branch] gh/janeyx99/329/head -> origin/gh/janeyx99/329/head 2025-12-04T09:20:54.3210034Z * [new branch] gh/janeyx99/329/orig -> origin/gh/janeyx99/329/orig 2025-12-04T09:20:54.3210185Z * [new branch] gh/janeyx99/330/base -> origin/gh/janeyx99/330/base 2025-12-04T09:20:54.3210901Z * [new branch] gh/janeyx99/330/head -> origin/gh/janeyx99/330/head 2025-12-04T09:20:54.3211101Z * [new branch] gh/janeyx99/330/orig -> origin/gh/janeyx99/330/orig 2025-12-04T09:20:54.3211276Z * [new branch] gh/janeyx99/331/base -> origin/gh/janeyx99/331/base 2025-12-04T09:20:54.3212248Z * [new branch] gh/janeyx99/331/head -> origin/gh/janeyx99/331/head 2025-12-04T09:20:54.3212705Z * [new branch] gh/janeyx99/331/orig -> origin/gh/janeyx99/331/orig 2025-12-04T09:20:54.3216773Z * [new branch] gh/janeyx99/332/base -> origin/gh/janeyx99/332/base 2025-12-04T09:20:54.3217287Z * [new branch] gh/janeyx99/332/head -> origin/gh/janeyx99/332/head 2025-12-04T09:20:54.3217436Z * [new branch] gh/janeyx99/332/orig -> origin/gh/janeyx99/332/orig 2025-12-04T09:20:54.3217606Z * [new branch] gh/janeyx99/333/base -> origin/gh/janeyx99/333/base 2025-12-04T09:20:54.3217749Z * [new branch] gh/janeyx99/333/head -> origin/gh/janeyx99/333/head 2025-12-04T09:20:54.3217900Z * [new branch] gh/janeyx99/333/orig -> origin/gh/janeyx99/333/orig 2025-12-04T09:20:54.3218986Z * [new branch] gh/janeyx99/88/base -> origin/gh/janeyx99/88/base 2025-12-04T09:20:54.3222993Z * [new branch] gh/janeyx99/88/head -> origin/gh/janeyx99/88/head 2025-12-04T09:20:54.3223198Z * [new branch] gh/janeyx99/88/orig -> origin/gh/janeyx99/88/orig 2025-12-04T09:20:54.3223375Z * [new branch] gh/jansel/360/base -> origin/gh/jansel/360/base 2025-12-04T09:20:54.3223530Z * [new branch] gh/jansel/360/head -> origin/gh/jansel/360/head 2025-12-04T09:20:54.3223670Z * [new branch] gh/jansel/451/base -> origin/gh/jansel/451/base 2025-12-04T09:20:54.3224180Z * [new branch] gh/jansel/451/head -> origin/gh/jansel/451/head 2025-12-04T09:20:54.3225007Z * [new branch] gh/jansel/451/orig -> origin/gh/jansel/451/orig 2025-12-04T09:20:54.3229227Z * [new branch] gh/jansel/462/base -> origin/gh/jansel/462/base 2025-12-04T09:20:54.3229433Z * [new branch] gh/jansel/462/head -> origin/gh/jansel/462/head 2025-12-04T09:20:54.3229574Z * [new branch] gh/jansel/462/orig -> origin/gh/jansel/462/orig 2025-12-04T09:20:54.3229710Z * [new branch] gh/jansel/533/base -> origin/gh/jansel/533/base 2025-12-04T09:20:54.3229878Z * [new branch] gh/jansel/533/head -> origin/gh/jansel/533/head 2025-12-04T09:20:54.3230015Z * [new branch] gh/jansel/533/orig -> origin/gh/jansel/533/orig 2025-12-04T09:20:54.3235250Z * [new branch] gh/jansel/552/base -> origin/gh/jansel/552/base 2025-12-04T09:20:54.3235443Z * [new branch] gh/jansel/552/head -> origin/gh/jansel/552/head 2025-12-04T09:20:54.3235602Z * [new branch] gh/jansel/552/orig -> origin/gh/jansel/552/orig 2025-12-04T09:20:54.3235749Z * [new branch] gh/jansel/553/base -> origin/gh/jansel/553/base 2025-12-04T09:20:54.3236271Z * [new branch] gh/jansel/553/head -> origin/gh/jansel/553/head 2025-12-04T09:20:54.3236899Z * [new branch] gh/jansel/553/orig -> origin/gh/jansel/553/orig 2025-12-04T09:20:54.3238203Z * [new branch] gh/jansel/554/base -> origin/gh/jansel/554/base 2025-12-04T09:20:54.3238618Z * [new branch] gh/jansel/554/head -> origin/gh/jansel/554/head 2025-12-04T09:20:54.3240864Z * [new branch] gh/jansel/554/orig -> origin/gh/jansel/554/orig 2025-12-04T09:20:54.3241063Z * [new branch] gh/jansel/555/base -> origin/gh/jansel/555/base 2025-12-04T09:20:54.3242736Z * [new branch] gh/jansel/555/head -> origin/gh/jansel/555/head 2025-12-04T09:20:54.3243038Z * [new branch] gh/jansel/555/orig -> origin/gh/jansel/555/orig 2025-12-04T09:20:54.3243338Z * [new branch] gh/jansel/556/base -> origin/gh/jansel/556/base 2025-12-04T09:20:54.3243554Z * [new branch] gh/jansel/556/head -> origin/gh/jansel/556/head 2025-12-04T09:20:54.3249133Z * [new branch] gh/jansel/556/orig -> origin/gh/jansel/556/orig 2025-12-04T09:20:54.3249457Z * [new branch] gh/jansel/557/base -> origin/gh/jansel/557/base 2025-12-04T09:20:54.3249645Z * [new branch] gh/jansel/557/head -> origin/gh/jansel/557/head 2025-12-04T09:20:54.3249792Z * [new branch] gh/jansel/557/orig -> origin/gh/jansel/557/orig 2025-12-04T09:20:54.3250066Z * [new branch] gh/jansel/558/base -> origin/gh/jansel/558/base 2025-12-04T09:20:54.3250309Z * [new branch] gh/jansel/558/head -> origin/gh/jansel/558/head 2025-12-04T09:20:54.3255058Z * [new branch] gh/jansel/558/orig -> origin/gh/jansel/558/orig 2025-12-04T09:20:54.3255576Z * [new branch] gh/jansel/559/base -> origin/gh/jansel/559/base 2025-12-04T09:20:54.3255762Z * [new branch] gh/jansel/559/head -> origin/gh/jansel/559/head 2025-12-04T09:20:54.3255958Z * [new branch] gh/jansel/559/orig -> origin/gh/jansel/559/orig 2025-12-04T09:20:54.3256119Z * [new branch] gh/jansel/560/base -> origin/gh/jansel/560/base 2025-12-04T09:20:54.3258915Z * [new branch] gh/jansel/560/head -> origin/gh/jansel/560/head 2025-12-04T09:20:54.3259112Z * [new branch] gh/jansel/560/orig -> origin/gh/jansel/560/orig 2025-12-04T09:20:54.3259271Z * [new branch] gh/jansel/561/base -> origin/gh/jansel/561/base 2025-12-04T09:20:54.3259453Z * [new branch] gh/jansel/561/head -> origin/gh/jansel/561/head 2025-12-04T09:20:54.3259604Z * [new branch] gh/jansel/561/orig -> origin/gh/jansel/561/orig 2025-12-04T09:20:54.3259749Z * [new branch] gh/jansel/562/base -> origin/gh/jansel/562/base 2025-12-04T09:20:54.3263969Z * [new branch] gh/jansel/562/head -> origin/gh/jansel/562/head 2025-12-04T09:20:54.3264164Z * [new branch] gh/jansel/562/orig -> origin/gh/jansel/562/orig 2025-12-04T09:20:54.3264573Z * [new branch] gh/jansel/563/base -> origin/gh/jansel/563/base 2025-12-04T09:20:54.3264783Z * [new branch] gh/jansel/563/head -> origin/gh/jansel/563/head 2025-12-04T09:20:54.3264932Z * [new branch] gh/jansel/563/orig -> origin/gh/jansel/563/orig 2025-12-04T09:20:54.3265552Z * [new branch] gh/jansel/564/base -> origin/gh/jansel/564/base 2025-12-04T09:20:54.3265703Z * [new branch] gh/jansel/564/head -> origin/gh/jansel/564/head 2025-12-04T09:20:54.3265846Z * [new branch] gh/jansel/564/orig -> origin/gh/jansel/564/orig 2025-12-04T09:20:54.3266136Z * [new branch] gh/jansel/565/base -> origin/gh/jansel/565/base 2025-12-04T09:20:54.3266369Z * [new branch] gh/jansel/565/head -> origin/gh/jansel/565/head 2025-12-04T09:20:54.3266535Z * [new branch] gh/jansel/565/orig -> origin/gh/jansel/565/orig 2025-12-04T09:20:54.3272788Z * [new branch] gh/jansel/566/base -> origin/gh/jansel/566/base 2025-12-04T09:20:54.3272990Z * [new branch] gh/jansel/566/head -> origin/gh/jansel/566/head 2025-12-04T09:20:54.3273135Z * [new branch] gh/jansel/566/orig -> origin/gh/jansel/566/orig 2025-12-04T09:20:54.3273281Z * [new branch] gh/jansel/567/base -> origin/gh/jansel/567/base 2025-12-04T09:20:54.3273838Z * [new branch] gh/jansel/567/head -> origin/gh/jansel/567/head 2025-12-04T09:20:54.3274220Z * [new branch] gh/jansel/567/orig -> origin/gh/jansel/567/orig 2025-12-04T09:20:54.3274491Z * [new branch] gh/jansel/568/base -> origin/gh/jansel/568/base 2025-12-04T09:20:54.3274644Z * [new branch] gh/jansel/568/head -> origin/gh/jansel/568/head 2025-12-04T09:20:54.3274794Z * [new branch] gh/jansel/568/orig -> origin/gh/jansel/568/orig 2025-12-04T09:20:54.3274944Z * [new branch] gh/jansel/569/base -> origin/gh/jansel/569/base 2025-12-04T09:20:54.3275092Z * [new branch] gh/jansel/569/head -> origin/gh/jansel/569/head 2025-12-04T09:20:54.3278804Z * [new branch] gh/jansel/569/orig -> origin/gh/jansel/569/orig 2025-12-04T09:20:54.3278971Z * [new branch] gh/jansel/570/base -> origin/gh/jansel/570/base 2025-12-04T09:20:54.3279628Z * [new branch] gh/jansel/570/head -> origin/gh/jansel/570/head 2025-12-04T09:20:54.3279937Z * [new branch] gh/jansel/570/orig -> origin/gh/jansel/570/orig 2025-12-04T09:20:54.3280264Z * [new branch] gh/jansel/571/base -> origin/gh/jansel/571/base 2025-12-04T09:20:54.3280428Z * [new branch] gh/jansel/571/head -> origin/gh/jansel/571/head 2025-12-04T09:20:54.3282433Z * [new branch] gh/jansel/571/orig -> origin/gh/jansel/571/orig 2025-12-04T09:20:54.3282721Z * [new branch] gh/jansel/572/base -> origin/gh/jansel/572/base 2025-12-04T09:20:54.3282874Z * [new branch] gh/jansel/572/head -> origin/gh/jansel/572/head 2025-12-04T09:20:54.3283024Z * [new branch] gh/jansel/572/orig -> origin/gh/jansel/572/orig 2025-12-04T09:20:54.3285098Z * [new branch] gh/jansel/573/base -> origin/gh/jansel/573/base 2025-12-04T09:20:54.3285289Z * [new branch] gh/jansel/573/head -> origin/gh/jansel/573/head 2025-12-04T09:20:54.3285435Z * [new branch] gh/jansel/573/orig -> origin/gh/jansel/573/orig 2025-12-04T09:20:54.3286389Z * [new branch] gh/jansel/574/base -> origin/gh/jansel/574/base 2025-12-04T09:20:54.3286675Z * [new branch] gh/jansel/574/head -> origin/gh/jansel/574/head 2025-12-04T09:20:54.3288114Z * [new branch] gh/jansel/574/orig -> origin/gh/jansel/574/orig 2025-12-04T09:20:54.3288415Z * [new branch] gh/jansel/575/base -> origin/gh/jansel/575/base 2025-12-04T09:20:54.3289980Z * [new branch] gh/jansel/575/head -> origin/gh/jansel/575/head 2025-12-04T09:20:54.3290396Z * [new branch] gh/jansel/575/orig -> origin/gh/jansel/575/orig 2025-12-04T09:20:54.3290840Z * [new branch] gh/jansel/576/base -> origin/gh/jansel/576/base 2025-12-04T09:20:54.3295774Z * [new branch] gh/jansel/576/head -> origin/gh/jansel/576/head 2025-12-04T09:20:54.3300920Z * [new branch] gh/jansel/576/orig -> origin/gh/jansel/576/orig 2025-12-04T09:20:54.3305586Z * [new branch] gh/jbschlosser/247/base -> origin/gh/jbschlosser/247/base 2025-12-04T09:20:54.3305791Z * [new branch] gh/jbschlosser/247/head -> origin/gh/jbschlosser/247/head 2025-12-04T09:20:54.3305986Z * [new branch] gh/jbschlosser/247/orig -> origin/gh/jbschlosser/247/orig 2025-12-04T09:20:54.3306342Z * [new branch] gh/jbschlosser/250/base -> origin/gh/jbschlosser/250/base 2025-12-04T09:20:54.3306500Z * [new branch] gh/jbschlosser/250/head -> origin/gh/jbschlosser/250/head 2025-12-04T09:20:54.3306669Z * [new branch] gh/jbschlosser/250/orig -> origin/gh/jbschlosser/250/orig 2025-12-04T09:20:54.3306879Z * [new branch] gh/jerryzh168/1/base -> origin/gh/jerryzh168/1/base 2025-12-04T09:20:54.3307035Z * [new branch] gh/jerryzh168/1/head -> origin/gh/jerryzh168/1/head 2025-12-04T09:20:54.3307192Z * [new branch] gh/jerryzh168/1/orig -> origin/gh/jerryzh168/1/orig 2025-12-04T09:20:54.3307341Z * [new branch] gh/jiayisunx/59/base -> origin/gh/jiayisunx/59/base 2025-12-04T09:20:54.3307495Z * [new branch] gh/jiayisunx/59/head -> origin/gh/jiayisunx/59/head 2025-12-04T09:20:54.3307639Z * [new branch] gh/jiayisunx/59/orig -> origin/gh/jiayisunx/59/orig 2025-12-04T09:20:54.3307786Z * [new branch] gh/jiayisunx/61/base -> origin/gh/jiayisunx/61/base 2025-12-04T09:20:54.3307938Z * [new branch] gh/jiayisunx/61/head -> origin/gh/jiayisunx/61/head 2025-12-04T09:20:54.3308084Z * [new branch] gh/jiayisunx/61/orig -> origin/gh/jiayisunx/61/orig 2025-12-04T09:20:54.3308243Z * [new branch] gh/jiayisunx/68/base -> origin/gh/jiayisunx/68/base 2025-12-04T09:20:54.3312312Z * [new branch] gh/jiayisunx/68/head -> origin/gh/jiayisunx/68/head 2025-12-04T09:20:54.3312501Z * [new branch] gh/jiayisunx/68/orig -> origin/gh/jiayisunx/68/orig 2025-12-04T09:20:54.3312667Z * [new branch] gh/jiayisunx/77/base -> origin/gh/jiayisunx/77/base 2025-12-04T09:20:54.3312815Z * [new branch] gh/jiayisunx/77/head -> origin/gh/jiayisunx/77/head 2025-12-04T09:20:54.3312988Z * [new branch] gh/jiayisunx/77/orig -> origin/gh/jiayisunx/77/orig 2025-12-04T09:20:54.3315013Z * [new branch] gh/jiayisunx/78/base -> origin/gh/jiayisunx/78/base 2025-12-04T09:20:54.3315366Z * [new branch] gh/jiayisunx/78/head -> origin/gh/jiayisunx/78/head 2025-12-04T09:20:54.3315550Z * [new branch] gh/jiayisunx/78/orig -> origin/gh/jiayisunx/78/orig 2025-12-04T09:20:54.3315785Z * [new branch] gh/jiayisunx/79/base -> origin/gh/jiayisunx/79/base 2025-12-04T09:20:54.3315946Z * [new branch] gh/jiayisunx/79/head -> origin/gh/jiayisunx/79/head 2025-12-04T09:20:54.3316113Z * [new branch] gh/jiayisunx/79/orig -> origin/gh/jiayisunx/79/orig 2025-12-04T09:20:54.3316275Z * [new branch] gh/jiayisunx/82/base -> origin/gh/jiayisunx/82/base 2025-12-04T09:20:54.3316434Z * [new branch] gh/jiayisunx/82/head -> origin/gh/jiayisunx/82/head 2025-12-04T09:20:54.3319744Z * [new branch] gh/jiayisunx/82/orig -> origin/gh/jiayisunx/82/orig 2025-12-04T09:20:54.3320411Z * [new branch] gh/jiayisunx/83/base -> origin/gh/jiayisunx/83/base 2025-12-04T09:20:54.3320602Z * [new branch] gh/jiayisunx/83/head -> origin/gh/jiayisunx/83/head 2025-12-04T09:20:54.3320780Z * [new branch] gh/jiayisunx/83/orig -> origin/gh/jiayisunx/83/orig 2025-12-04T09:20:54.3320930Z * [new branch] gh/jiayisunx/84/base -> origin/gh/jiayisunx/84/base 2025-12-04T09:20:54.3329219Z * [new branch] gh/jiayisunx/84/head -> origin/gh/jiayisunx/84/head 2025-12-04T09:20:54.3329425Z * [new branch] gh/jiayisunx/84/orig -> origin/gh/jiayisunx/84/orig 2025-12-04T09:20:54.3329673Z * [new branch] gh/jiayisunx/85/base -> origin/gh/jiayisunx/85/base 2025-12-04T09:20:54.3329852Z * [new branch] gh/jiayisunx/85/head -> origin/gh/jiayisunx/85/head 2025-12-04T09:20:54.3330532Z * [new branch] gh/jiayisunx/85/orig -> origin/gh/jiayisunx/85/orig 2025-12-04T09:20:54.3330715Z * [new branch] gh/jiayisunx/86/base -> origin/gh/jiayisunx/86/base 2025-12-04T09:20:54.3331351Z * [new branch] gh/jiayisunx/86/head -> origin/gh/jiayisunx/86/head 2025-12-04T09:20:54.3331860Z * [new branch] gh/jiayisunx/86/orig -> origin/gh/jiayisunx/86/orig 2025-12-04T09:20:54.3332026Z * [new branch] gh/jiayisunx/87/base -> origin/gh/jiayisunx/87/base 2025-12-04T09:20:54.3332181Z * [new branch] gh/jiayisunx/87/head -> origin/gh/jiayisunx/87/head 2025-12-04T09:20:54.3332348Z * [new branch] gh/jiayisunx/87/orig -> origin/gh/jiayisunx/87/orig 2025-12-04T09:20:54.3332500Z * [new branch] gh/jiayisunx/88/base -> origin/gh/jiayisunx/88/base 2025-12-04T09:20:54.3332654Z * [new branch] gh/jiayisunx/88/head -> origin/gh/jiayisunx/88/head 2025-12-04T09:20:54.3332821Z * [new branch] gh/jiayisunx/88/orig -> origin/gh/jiayisunx/88/orig 2025-12-04T09:20:54.3342134Z * [new branch] gh/jiayisunx/89/base -> origin/gh/jiayisunx/89/base 2025-12-04T09:20:54.3345377Z * [new branch] gh/jiayisunx/89/head -> origin/gh/jiayisunx/89/head 2025-12-04T09:20:54.3345695Z * [new branch] gh/jiayisunx/89/orig -> origin/gh/jiayisunx/89/orig 2025-12-04T09:20:54.3345874Z * [new branch] gh/jiayisunx/90/base -> origin/gh/jiayisunx/90/base 2025-12-04T09:20:54.3346033Z * [new branch] gh/jiayisunx/90/head -> origin/gh/jiayisunx/90/head 2025-12-04T09:20:54.3346197Z * [new branch] gh/jiayisunx/90/orig -> origin/gh/jiayisunx/90/orig 2025-12-04T09:20:54.3346386Z * [new branch] gh/jjwu@meta.com/1/base -> origin/gh/jjwu@meta.com/1/base 2025-12-04T09:20:54.3346571Z * [new branch] gh/jjwu@meta.com/1/head -> origin/gh/jjwu@meta.com/1/head 2025-12-04T09:20:54.3349794Z * [new branch] gh/jturney/1/base -> origin/gh/jturney/1/base 2025-12-04T09:20:54.3349977Z * [new branch] gh/jturney/1/head -> origin/gh/jturney/1/head 2025-12-04T09:20:54.3350149Z * [new branch] gh/jturney/1/orig -> origin/gh/jturney/1/orig 2025-12-04T09:20:54.3350292Z * [new branch] gh/jturney/2/base -> origin/gh/jturney/2/base 2025-12-04T09:20:54.3350435Z * [new branch] gh/jturney/2/head -> origin/gh/jturney/2/head 2025-12-04T09:20:54.3356304Z * [new branch] gh/jturney/2/orig -> origin/gh/jturney/2/orig 2025-12-04T09:20:54.3356510Z * [new branch] gh/karthickai/10/base -> origin/gh/karthickai/10/base 2025-12-04T09:20:54.3356680Z * [new branch] gh/karthickai/10/head -> origin/gh/karthickai/10/head 2025-12-04T09:20:54.3356856Z * [new branch] gh/karthickai/10/orig -> origin/gh/karthickai/10/orig 2025-12-04T09:20:54.3357021Z * [new branch] gh/karthickai/11/base -> origin/gh/karthickai/11/base 2025-12-04T09:20:54.3357932Z * [new branch] gh/karthickai/11/head -> origin/gh/karthickai/11/head 2025-12-04T09:20:54.3358282Z * [new branch] gh/karthickai/11/orig -> origin/gh/karthickai/11/orig 2025-12-04T09:20:54.3358492Z * [new branch] gh/karthickai/12/base -> origin/gh/karthickai/12/base 2025-12-04T09:20:54.3358662Z * [new branch] gh/karthickai/12/head -> origin/gh/karthickai/12/head 2025-12-04T09:20:54.3358821Z * [new branch] gh/karthickai/12/orig -> origin/gh/karthickai/12/orig 2025-12-04T09:20:54.3358981Z * [new branch] gh/karthickai/13/base -> origin/gh/karthickai/13/base 2025-12-04T09:20:54.3359142Z * [new branch] gh/karthickai/13/head -> origin/gh/karthickai/13/head 2025-12-04T09:20:54.3359644Z * [new branch] gh/karthickai/13/orig -> origin/gh/karthickai/13/orig 2025-12-04T09:20:54.3361508Z * [new branch] gh/karthickai/14/base -> origin/gh/karthickai/14/base 2025-12-04T09:20:54.3361748Z * [new branch] gh/karthickai/14/head -> origin/gh/karthickai/14/head 2025-12-04T09:20:54.3365462Z * [new branch] gh/karthickai/14/orig -> origin/gh/karthickai/14/orig 2025-12-04T09:20:54.3365884Z * [new branch] gh/karthickai/15/base -> origin/gh/karthickai/15/base 2025-12-04T09:20:54.3366051Z * [new branch] gh/karthickai/15/head -> origin/gh/karthickai/15/head 2025-12-04T09:20:54.3366211Z * [new branch] gh/karthickai/15/orig -> origin/gh/karthickai/15/orig 2025-12-04T09:20:54.3366371Z * [new branch] gh/karthickai/16/base -> origin/gh/karthickai/16/base 2025-12-04T09:20:54.3366906Z * [new branch] gh/karthickai/16/head -> origin/gh/karthickai/16/head 2025-12-04T09:20:54.3372614Z * [new branch] gh/karthickai/16/orig -> origin/gh/karthickai/16/orig 2025-12-04T09:20:54.3372839Z * [new branch] gh/karthickai/17/base -> origin/gh/karthickai/17/base 2025-12-04T09:20:54.3373006Z * [new branch] gh/karthickai/17/head -> origin/gh/karthickai/17/head 2025-12-04T09:20:54.3373190Z * [new branch] gh/karthickai/17/orig -> origin/gh/karthickai/17/orig 2025-12-04T09:20:54.3373354Z * [new branch] gh/karthickai/18/base -> origin/gh/karthickai/18/base 2025-12-04T09:20:54.3373522Z * [new branch] gh/karthickai/18/head -> origin/gh/karthickai/18/head 2025-12-04T09:20:54.3373683Z * [new branch] gh/karthickai/18/orig -> origin/gh/karthickai/18/orig 2025-12-04T09:20:54.3374117Z * [new branch] gh/karthickai/19/base -> origin/gh/karthickai/19/base 2025-12-04T09:20:54.3374517Z * [new branch] gh/karthickai/19/head -> origin/gh/karthickai/19/head 2025-12-04T09:20:54.3374744Z * [new branch] gh/karthickai/19/orig -> origin/gh/karthickai/19/orig 2025-12-04T09:20:54.3381258Z * [new branch] gh/karthickai/20/base -> origin/gh/karthickai/20/base 2025-12-04T09:20:54.3383698Z * [new branch] gh/karthickai/20/head -> origin/gh/karthickai/20/head 2025-12-04T09:20:54.3383999Z * [new branch] gh/karthickai/20/orig -> origin/gh/karthickai/20/orig 2025-12-04T09:20:54.3386920Z * [new branch] gh/karthickai/21/base -> origin/gh/karthickai/21/base 2025-12-04T09:20:54.3387221Z * [new branch] gh/karthickai/21/head -> origin/gh/karthickai/21/head 2025-12-04T09:20:54.3391362Z * [new branch] gh/karthickai/21/orig -> origin/gh/karthickai/21/orig 2025-12-04T09:20:54.3395804Z * [new branch] gh/karthickai/22/base -> origin/gh/karthickai/22/base 2025-12-04T09:20:54.3398331Z * [new branch] gh/karthickai/22/head -> origin/gh/karthickai/22/head 2025-12-04T09:20:54.3398927Z * [new branch] gh/karthickai/22/orig -> origin/gh/karthickai/22/orig 2025-12-04T09:20:54.3399116Z * [new branch] gh/karthickai/23/base -> origin/gh/karthickai/23/base 2025-12-04T09:20:54.3399300Z * [new branch] gh/karthickai/23/head -> origin/gh/karthickai/23/head 2025-12-04T09:20:54.3399462Z * [new branch] gh/karthickai/23/orig -> origin/gh/karthickai/23/orig 2025-12-04T09:20:54.3399612Z * [new branch] gh/karthickai/24/base -> origin/gh/karthickai/24/base 2025-12-04T09:20:54.3399764Z * [new branch] gh/karthickai/24/head -> origin/gh/karthickai/24/head 2025-12-04T09:20:54.3399922Z * [new branch] gh/karthickai/24/orig -> origin/gh/karthickai/24/orig 2025-12-04T09:20:54.3400070Z * [new branch] gh/karthickai/25/base -> origin/gh/karthickai/25/base 2025-12-04T09:20:54.3400427Z * [new branch] gh/karthickai/25/head -> origin/gh/karthickai/25/head 2025-12-04T09:20:54.3400627Z * [new branch] gh/karthickai/25/orig -> origin/gh/karthickai/25/orig 2025-12-04T09:20:54.3400801Z * [new branch] gh/karthickai/26/base -> origin/gh/karthickai/26/base 2025-12-04T09:20:54.3401008Z * [new branch] gh/karthickai/26/head -> origin/gh/karthickai/26/head 2025-12-04T09:20:54.3401158Z * [new branch] gh/karthickai/26/orig -> origin/gh/karthickai/26/orig 2025-12-04T09:20:54.3401334Z * [new branch] gh/karthickai/6/base -> origin/gh/karthickai/6/base 2025-12-04T09:20:54.3401492Z * [new branch] gh/karthickai/6/head -> origin/gh/karthickai/6/head 2025-12-04T09:20:54.3401645Z * [new branch] gh/karthickai/6/orig -> origin/gh/karthickai/6/orig 2025-12-04T09:20:54.3401876Z * [new branch] gh/krocki/1/base -> origin/gh/krocki/1/base 2025-12-04T09:20:54.3402038Z * [new branch] gh/krocki/1/head -> origin/gh/krocki/1/head 2025-12-04T09:20:54.3402227Z * [new branch] gh/krocki/1/orig -> origin/gh/krocki/1/orig 2025-12-04T09:20:54.3410267Z * [new branch] gh/krocki/2/base -> origin/gh/krocki/2/base 2025-12-04T09:20:54.3415276Z * [new branch] gh/krocki/2/head -> origin/gh/krocki/2/head 2025-12-04T09:20:54.3417190Z * [new branch] gh/krocki/2/orig -> origin/gh/krocki/2/orig 2025-12-04T09:20:54.3417512Z * [new branch] gh/kurtamohler/60/base -> origin/gh/kurtamohler/60/base 2025-12-04T09:20:54.3425550Z * [new branch] gh/kurtamohler/60/head -> origin/gh/kurtamohler/60/head 2025-12-04T09:20:54.3426074Z * [new branch] gh/kurtamohler/60/orig -> origin/gh/kurtamohler/60/orig 2025-12-04T09:20:54.3426279Z * [new branch] gh/kurtamohler/61/base -> origin/gh/kurtamohler/61/base 2025-12-04T09:20:54.3426439Z * [new branch] gh/kurtamohler/61/head -> origin/gh/kurtamohler/61/head 2025-12-04T09:20:54.3426601Z * [new branch] gh/kurtamohler/61/orig -> origin/gh/kurtamohler/61/orig 2025-12-04T09:20:54.3426761Z * [new branch] gh/kurtamohler/62/base -> origin/gh/kurtamohler/62/base 2025-12-04T09:20:54.3426924Z * [new branch] gh/kurtamohler/62/head -> origin/gh/kurtamohler/62/head 2025-12-04T09:20:54.3427077Z * [new branch] gh/kurtamohler/62/orig -> origin/gh/kurtamohler/62/orig 2025-12-04T09:20:54.3427231Z * [new branch] gh/kurtamohler/63/base -> origin/gh/kurtamohler/63/base 2025-12-04T09:20:54.3427389Z * [new branch] gh/kurtamohler/63/head -> origin/gh/kurtamohler/63/head 2025-12-04T09:20:54.3427543Z * [new branch] gh/kurtamohler/63/orig -> origin/gh/kurtamohler/63/orig 2025-12-04T09:20:54.3427700Z * [new branch] gh/kurtamohler/64/base -> origin/gh/kurtamohler/64/base 2025-12-04T09:20:54.3427902Z * [new branch] gh/kurtamohler/64/head -> origin/gh/kurtamohler/64/head 2025-12-04T09:20:54.3428061Z * [new branch] gh/kurtamohler/64/orig -> origin/gh/kurtamohler/64/orig 2025-12-04T09:20:54.3428224Z * [new branch] gh/kurtamohler/65/base -> origin/gh/kurtamohler/65/base 2025-12-04T09:20:54.3428375Z * [new branch] gh/kurtamohler/65/head -> origin/gh/kurtamohler/65/head 2025-12-04T09:20:54.3428528Z * [new branch] gh/kurtamohler/65/orig -> origin/gh/kurtamohler/65/orig 2025-12-04T09:20:54.3428689Z * [new branch] gh/kurtamohler/66/base -> origin/gh/kurtamohler/66/base 2025-12-04T09:20:54.3428839Z * [new branch] gh/kurtamohler/66/head -> origin/gh/kurtamohler/66/head 2025-12-04T09:20:54.3429142Z * [new branch] gh/kurtamohler/66/orig -> origin/gh/kurtamohler/66/orig 2025-12-04T09:20:54.3429294Z * [new branch] gh/kurtamohler/67/base -> origin/gh/kurtamohler/67/base 2025-12-04T09:20:54.3429477Z * [new branch] gh/kurtamohler/67/head -> origin/gh/kurtamohler/67/head 2025-12-04T09:20:54.3429691Z * [new branch] gh/kurtamohler/67/orig -> origin/gh/kurtamohler/67/orig 2025-12-04T09:20:54.3429854Z * [new branch] gh/kwen2501/130/base -> origin/gh/kwen2501/130/base 2025-12-04T09:20:54.3430018Z * [new branch] gh/kwen2501/130/head -> origin/gh/kwen2501/130/head 2025-12-04T09:20:54.3430290Z * [new branch] gh/kwen2501/130/orig -> origin/gh/kwen2501/130/orig 2025-12-04T09:20:54.3430440Z * [new branch] gh/kwen2501/170/base -> origin/gh/kwen2501/170/base 2025-12-04T09:20:54.3430587Z * [new branch] gh/kwen2501/170/head -> origin/gh/kwen2501/170/head 2025-12-04T09:20:54.3430730Z * [new branch] gh/kwen2501/187/base -> origin/gh/kwen2501/187/base 2025-12-04T09:20:54.3431083Z * [new branch] gh/kwen2501/187/head -> origin/gh/kwen2501/187/head 2025-12-04T09:20:54.3431545Z * [new branch] gh/kwen2501/187/orig -> origin/gh/kwen2501/187/orig 2025-12-04T09:20:54.3435875Z * [new branch] gh/kwen2501/188/base -> origin/gh/kwen2501/188/base 2025-12-04T09:20:54.3436069Z * [new branch] gh/kwen2501/188/head -> origin/gh/kwen2501/188/head 2025-12-04T09:20:54.3436228Z * [new branch] gh/kwen2501/188/orig -> origin/gh/kwen2501/188/orig 2025-12-04T09:20:54.3436387Z * [new branch] gh/kwen2501/211/base -> origin/gh/kwen2501/211/base 2025-12-04T09:20:54.3436549Z * [new branch] gh/kwen2501/211/head -> origin/gh/kwen2501/211/head 2025-12-04T09:20:54.3437740Z * [new branch] gh/kwen2501/224/base -> origin/gh/kwen2501/224/base 2025-12-04T09:20:54.3437918Z * [new branch] gh/kwen2501/224/head -> origin/gh/kwen2501/224/head 2025-12-04T09:20:54.3438088Z * [new branch] gh/kwen2501/224/orig -> origin/gh/kwen2501/224/orig 2025-12-04T09:20:54.3442224Z * [new branch] gh/kwen2501/228/base -> origin/gh/kwen2501/228/base 2025-12-04T09:20:54.3442435Z * [new branch] gh/kwen2501/228/head -> origin/gh/kwen2501/228/head 2025-12-04T09:20:54.3442594Z * [new branch] gh/kwen2501/228/orig -> origin/gh/kwen2501/228/orig 2025-12-04T09:20:54.3442749Z * [new branch] gh/kwen2501/234/base -> origin/gh/kwen2501/234/base 2025-12-04T09:20:54.3442910Z * [new branch] gh/kwen2501/234/head -> origin/gh/kwen2501/234/head 2025-12-04T09:20:54.3447415Z * [new branch] gh/kwen2501/234/orig -> origin/gh/kwen2501/234/orig 2025-12-04T09:20:54.3452639Z * [new branch] gh/kwen2501/235/base -> origin/gh/kwen2501/235/base 2025-12-04T09:20:54.3453678Z * [new branch] gh/kwen2501/235/head -> origin/gh/kwen2501/235/head 2025-12-04T09:20:54.3453847Z * [new branch] gh/kwen2501/235/orig -> origin/gh/kwen2501/235/orig 2025-12-04T09:20:54.3453994Z * [new branch] gh/kwen2501/236/base -> origin/gh/kwen2501/236/base 2025-12-04T09:20:54.3454145Z * [new branch] gh/kwen2501/236/head -> origin/gh/kwen2501/236/head 2025-12-04T09:20:54.3454296Z * [new branch] gh/kwen2501/236/orig -> origin/gh/kwen2501/236/orig 2025-12-04T09:20:54.3454438Z * [new branch] gh/kwen2501/237/base -> origin/gh/kwen2501/237/base 2025-12-04T09:20:54.3454591Z * [new branch] gh/kwen2501/237/head -> origin/gh/kwen2501/237/head 2025-12-04T09:20:54.3454734Z * [new branch] gh/kwen2501/237/orig -> origin/gh/kwen2501/237/orig 2025-12-04T09:20:54.3455099Z * [new branch] gh/kwen2501/238/base -> origin/gh/kwen2501/238/base 2025-12-04T09:20:54.3455248Z * [new branch] gh/kwen2501/238/head -> origin/gh/kwen2501/238/head 2025-12-04T09:20:54.3455391Z * [new branch] gh/kwen2501/238/orig -> origin/gh/kwen2501/238/orig 2025-12-04T09:20:54.3455628Z * [new branch] gh/kwen2501/240/base -> origin/gh/kwen2501/240/base 2025-12-04T09:20:54.3457390Z * [new branch] gh/kwen2501/240/head -> origin/gh/kwen2501/240/head 2025-12-04T09:20:54.3458075Z * [new branch] gh/kwen2501/240/orig -> origin/gh/kwen2501/240/orig 2025-12-04T09:20:54.3461594Z * [new branch] gh/kwen2501/241/base -> origin/gh/kwen2501/241/base 2025-12-04T09:20:54.3461783Z * [new branch] gh/kwen2501/241/head -> origin/gh/kwen2501/241/head 2025-12-04T09:20:54.3461955Z * [new branch] gh/kwen2501/241/orig -> origin/gh/kwen2501/241/orig 2025-12-04T09:20:54.3462122Z * [new branch] gh/kwen2501/247/base -> origin/gh/kwen2501/247/base 2025-12-04T09:20:54.3462274Z * [new branch] gh/kwen2501/247/head -> origin/gh/kwen2501/247/head 2025-12-04T09:20:54.3462423Z * [new branch] gh/kwen2501/247/orig -> origin/gh/kwen2501/247/orig 2025-12-04T09:20:54.3462577Z * [new branch] gh/kwen2501/252/base -> origin/gh/kwen2501/252/base 2025-12-04T09:20:54.3462729Z * [new branch] gh/kwen2501/252/head -> origin/gh/kwen2501/252/head 2025-12-04T09:20:54.3462875Z * [new branch] gh/kwen2501/252/orig -> origin/gh/kwen2501/252/orig 2025-12-04T09:20:54.3467472Z * [new branch] gh/kwen2501/259/base -> origin/gh/kwen2501/259/base 2025-12-04T09:20:54.3467661Z * [new branch] gh/kwen2501/259/head -> origin/gh/kwen2501/259/head 2025-12-04T09:20:54.3467814Z * [new branch] gh/kwen2501/259/orig -> origin/gh/kwen2501/259/orig 2025-12-04T09:20:54.3467988Z * [new branch] gh/kwen2501/260/base -> origin/gh/kwen2501/260/base 2025-12-04T09:20:54.3468139Z * [new branch] gh/kwen2501/260/head -> origin/gh/kwen2501/260/head 2025-12-04T09:20:54.3468698Z * [new branch] gh/kwen2501/260/orig -> origin/gh/kwen2501/260/orig 2025-12-04T09:20:54.3470295Z * [new branch] gh/kwen2501/268/base -> origin/gh/kwen2501/268/base 2025-12-04T09:20:54.3470455Z * [new branch] gh/kwen2501/268/head -> origin/gh/kwen2501/268/head 2025-12-04T09:20:54.3471027Z * [new branch] gh/kwen2501/268/orig -> origin/gh/kwen2501/268/orig 2025-12-04T09:20:54.3474399Z * [new branch] gh/kwen2501/269/base -> origin/gh/kwen2501/269/base 2025-12-04T09:20:54.3474746Z * [new branch] gh/kwen2501/269/head -> origin/gh/kwen2501/269/head 2025-12-04T09:20:54.3475001Z * [new branch] gh/kwen2501/269/orig -> origin/gh/kwen2501/269/orig 2025-12-04T09:20:54.3475261Z * [new branch] gh/kwen2501/270/base -> origin/gh/kwen2501/270/base 2025-12-04T09:20:54.3475831Z * [new branch] gh/kwen2501/270/head -> origin/gh/kwen2501/270/head 2025-12-04T09:20:54.3477309Z * [new branch] gh/kwen2501/270/orig -> origin/gh/kwen2501/270/orig 2025-12-04T09:20:54.3477899Z * [new branch] gh/kwen2501/271/base -> origin/gh/kwen2501/271/base 2025-12-04T09:20:54.3479094Z * [new branch] gh/kwen2501/271/head -> origin/gh/kwen2501/271/head 2025-12-04T09:20:54.3479542Z * [new branch] gh/kwen2501/271/orig -> origin/gh/kwen2501/271/orig 2025-12-04T09:20:54.3481648Z * [new branch] gh/kwen2501/274/base -> origin/gh/kwen2501/274/base 2025-12-04T09:20:54.3482322Z * [new branch] gh/kwen2501/274/head -> origin/gh/kwen2501/274/head 2025-12-04T09:20:54.3482648Z * [new branch] gh/kwen2501/274/orig -> origin/gh/kwen2501/274/orig 2025-12-04T09:20:54.3484306Z * [new branch] gh/kwen2501/275/base -> origin/gh/kwen2501/275/base 2025-12-04T09:20:54.3484478Z * [new branch] gh/kwen2501/275/head -> origin/gh/kwen2501/275/head 2025-12-04T09:20:54.3484925Z * [new branch] gh/kwen2501/275/orig -> origin/gh/kwen2501/275/orig 2025-12-04T09:20:54.3486764Z * [new branch] gh/kwen2501/276/base -> origin/gh/kwen2501/276/base 2025-12-04T09:20:54.3487193Z * [new branch] gh/kwen2501/276/head -> origin/gh/kwen2501/276/head 2025-12-04T09:20:54.3487355Z * [new branch] gh/kwen2501/276/orig -> origin/gh/kwen2501/276/orig 2025-12-04T09:20:54.3490781Z * [new branch] gh/kwen2501/277/base -> origin/gh/kwen2501/277/base 2025-12-04T09:20:54.3490981Z * [new branch] gh/kwen2501/277/head -> origin/gh/kwen2501/277/head 2025-12-04T09:20:54.3491157Z * [new branch] gh/kwen2501/277/orig -> origin/gh/kwen2501/277/orig 2025-12-04T09:20:54.3491303Z * [new branch] gh/kwen2501/278/base -> origin/gh/kwen2501/278/base 2025-12-04T09:20:54.3491591Z * [new branch] gh/kwen2501/278/head -> origin/gh/kwen2501/278/head 2025-12-04T09:20:54.3494789Z * [new branch] gh/kwen2501/278/orig -> origin/gh/kwen2501/278/orig 2025-12-04T09:20:54.3495153Z * [new branch] gh/kwen2501/279/base -> origin/gh/kwen2501/279/base 2025-12-04T09:20:54.3495418Z * [new branch] gh/kwen2501/279/head -> origin/gh/kwen2501/279/head 2025-12-04T09:20:54.3495588Z * [new branch] gh/kwen2501/279/orig -> origin/gh/kwen2501/279/orig 2025-12-04T09:20:54.3499331Z * [new branch] gh/kwen2501/280/base -> origin/gh/kwen2501/280/base 2025-12-04T09:20:54.3499527Z * [new branch] gh/kwen2501/280/head -> origin/gh/kwen2501/280/head 2025-12-04T09:20:54.3500123Z * [new branch] gh/kwen2501/280/orig -> origin/gh/kwen2501/280/orig 2025-12-04T09:20:54.3500275Z * [new branch] gh/kwen2501/281/base -> origin/gh/kwen2501/281/base 2025-12-04T09:20:54.3500545Z * [new branch] gh/kwen2501/281/head -> origin/gh/kwen2501/281/head 2025-12-04T09:20:54.3506469Z * [new branch] gh/kwen2501/281/orig -> origin/gh/kwen2501/281/orig 2025-12-04T09:20:54.3506676Z * [new branch] gh/kwen2501/282/base -> origin/gh/kwen2501/282/base 2025-12-04T09:20:54.3506821Z * [new branch] gh/kwen2501/282/head -> origin/gh/kwen2501/282/head 2025-12-04T09:20:54.3507186Z * [new branch] gh/kwen2501/282/orig -> origin/gh/kwen2501/282/orig 2025-12-04T09:20:54.3507433Z * [new branch] gh/kwen2501/283/base -> origin/gh/kwen2501/283/base 2025-12-04T09:20:54.3507696Z * [new branch] gh/kwen2501/283/head -> origin/gh/kwen2501/283/head 2025-12-04T09:20:54.3507960Z * [new branch] gh/kwen2501/283/orig -> origin/gh/kwen2501/283/orig 2025-12-04T09:20:54.3508637Z * [new branch] gh/kwen2501/284/base -> origin/gh/kwen2501/284/base 2025-12-04T09:20:54.3508823Z * [new branch] gh/kwen2501/284/head -> origin/gh/kwen2501/284/head 2025-12-04T09:20:54.3546669Z * [new branch] gh/kwen2501/284/orig -> origin/gh/kwen2501/284/orig 2025-12-04T09:20:54.3547177Z * [new branch] gh/kwen2501/285/base -> origin/gh/kwen2501/285/base 2025-12-04T09:20:54.3547346Z * [new branch] gh/kwen2501/285/head -> origin/gh/kwen2501/285/head 2025-12-04T09:20:54.3547494Z * [new branch] gh/kwen2501/285/orig -> origin/gh/kwen2501/285/orig 2025-12-04T09:20:54.3547647Z * [new branch] gh/kwen2501/286/base -> origin/gh/kwen2501/286/base 2025-12-04T09:20:54.3548077Z * [new branch] gh/kwen2501/286/head -> origin/gh/kwen2501/286/head 2025-12-04T09:20:54.3548220Z * [new branch] gh/kwen2501/286/orig -> origin/gh/kwen2501/286/orig 2025-12-04T09:20:54.3548372Z * [new branch] gh/kwen2501/287/base -> origin/gh/kwen2501/287/base 2025-12-04T09:20:54.3548596Z * [new branch] gh/kwen2501/287/head -> origin/gh/kwen2501/287/head 2025-12-04T09:20:54.3548741Z * [new branch] gh/kwen2501/287/orig -> origin/gh/kwen2501/287/orig 2025-12-04T09:20:54.3548889Z * [new branch] gh/kwen2501/288/base -> origin/gh/kwen2501/288/base 2025-12-04T09:20:54.3549030Z * [new branch] gh/kwen2501/288/head -> origin/gh/kwen2501/288/head 2025-12-04T09:20:54.3549179Z * [new branch] gh/kwen2501/288/orig -> origin/gh/kwen2501/288/orig 2025-12-04T09:20:54.3549343Z * [new branch] gh/laithsakka/251/base -> origin/gh/laithsakka/251/base 2025-12-04T09:20:54.3549502Z * [new branch] gh/laithsakka/251/head -> origin/gh/laithsakka/251/head 2025-12-04T09:20:54.3549663Z * [new branch] gh/laithsakka/251/orig -> origin/gh/laithsakka/251/orig 2025-12-04T09:20:54.3549815Z * [new branch] gh/laithsakka/276/base -> origin/gh/laithsakka/276/base 2025-12-04T09:20:54.3549976Z * [new branch] gh/laithsakka/276/head -> origin/gh/laithsakka/276/head 2025-12-04T09:20:54.3550126Z * [new branch] gh/laithsakka/276/orig -> origin/gh/laithsakka/276/orig 2025-12-04T09:20:54.3550283Z * [new branch] gh/laithsakka/28/base -> origin/gh/laithsakka/28/base 2025-12-04T09:20:54.3550443Z * [new branch] gh/laithsakka/29/base -> origin/gh/laithsakka/29/base 2025-12-04T09:20:54.3550590Z * [new branch] gh/laithsakka/30/base -> origin/gh/laithsakka/30/base 2025-12-04T09:20:54.3550750Z * [new branch] gh/laithsakka/30/head -> origin/gh/laithsakka/30/head 2025-12-04T09:20:54.3550903Z * [new branch] gh/laithsakka/31/base -> origin/gh/laithsakka/31/base 2025-12-04T09:20:54.3551047Z * [new branch] gh/laithsakka/31/head -> origin/gh/laithsakka/31/head 2025-12-04T09:20:54.3551207Z * [new branch] gh/laithsakka/313/base -> origin/gh/laithsakka/313/base 2025-12-04T09:20:54.3551362Z * [new branch] gh/laithsakka/313/head -> origin/gh/laithsakka/313/head 2025-12-04T09:20:54.3551513Z * [new branch] gh/laithsakka/313/orig -> origin/gh/laithsakka/313/orig 2025-12-04T09:20:54.3551670Z * [new branch] gh/laithsakka/316/base -> origin/gh/laithsakka/316/base 2025-12-04T09:20:54.3551822Z * [new branch] gh/laithsakka/316/head -> origin/gh/laithsakka/316/head 2025-12-04T09:20:54.3551982Z * [new branch] gh/laithsakka/316/orig -> origin/gh/laithsakka/316/orig 2025-12-04T09:20:54.3552137Z * [new branch] gh/laithsakka/317/base -> origin/gh/laithsakka/317/base 2025-12-04T09:20:54.3552287Z * [new branch] gh/laithsakka/317/head -> origin/gh/laithsakka/317/head 2025-12-04T09:20:54.3552449Z * [new branch] gh/laithsakka/317/orig -> origin/gh/laithsakka/317/orig 2025-12-04T09:20:54.3552601Z * [new branch] gh/laithsakka/319/base -> origin/gh/laithsakka/319/base 2025-12-04T09:20:54.3552760Z * [new branch] gh/laithsakka/319/head -> origin/gh/laithsakka/319/head 2025-12-04T09:20:54.3552909Z * [new branch] gh/laithsakka/319/orig -> origin/gh/laithsakka/319/orig 2025-12-04T09:20:54.3553060Z * [new branch] gh/laithsakka/32/base -> origin/gh/laithsakka/32/base 2025-12-04T09:20:54.3553218Z * [new branch] gh/laithsakka/32/head -> origin/gh/laithsakka/32/head 2025-12-04T09:20:54.3553367Z * [new branch] gh/laithsakka/320/base -> origin/gh/laithsakka/320/base 2025-12-04T09:20:54.3553566Z * [new branch] gh/laithsakka/320/head -> origin/gh/laithsakka/320/head 2025-12-04T09:20:54.3553715Z * [new branch] gh/laithsakka/320/orig -> origin/gh/laithsakka/320/orig 2025-12-04T09:20:54.3553863Z * [new branch] gh/laithsakka/321/base -> origin/gh/laithsakka/321/base 2025-12-04T09:20:54.3554073Z * [new branch] gh/laithsakka/321/head -> origin/gh/laithsakka/321/head 2025-12-04T09:20:54.3554224Z * [new branch] gh/laithsakka/321/orig -> origin/gh/laithsakka/321/orig 2025-12-04T09:20:54.3554379Z * [new branch] gh/laithsakka/322/base -> origin/gh/laithsakka/322/base 2025-12-04T09:20:54.3554528Z * [new branch] gh/laithsakka/322/head -> origin/gh/laithsakka/322/head 2025-12-04T09:20:54.3554677Z * [new branch] gh/laithsakka/322/orig -> origin/gh/laithsakka/322/orig 2025-12-04T09:20:54.3554835Z * [new branch] gh/laithsakka/323/base -> origin/gh/laithsakka/323/base 2025-12-04T09:20:54.3554993Z * [new branch] gh/laithsakka/323/head -> origin/gh/laithsakka/323/head 2025-12-04T09:20:54.3555148Z * [new branch] gh/laithsakka/323/orig -> origin/gh/laithsakka/323/orig 2025-12-04T09:20:54.3555299Z * [new branch] gh/laithsakka/324/base -> origin/gh/laithsakka/324/base 2025-12-04T09:20:54.3555454Z * [new branch] gh/laithsakka/324/head -> origin/gh/laithsakka/324/head 2025-12-04T09:20:54.3555610Z * [new branch] gh/laithsakka/324/orig -> origin/gh/laithsakka/324/orig 2025-12-04T09:20:54.3555759Z * [new branch] gh/laithsakka/325/base -> origin/gh/laithsakka/325/base 2025-12-04T09:20:54.3555908Z * [new branch] gh/laithsakka/325/head -> origin/gh/laithsakka/325/head 2025-12-04T09:20:54.3556298Z * [new branch] gh/laithsakka/325/orig -> origin/gh/laithsakka/325/orig 2025-12-04T09:20:54.3557146Z * [new branch] gh/laithsakka/326/base -> origin/gh/laithsakka/326/base 2025-12-04T09:20:54.3557343Z * [new branch] gh/laithsakka/326/head -> origin/gh/laithsakka/326/head 2025-12-04T09:20:54.3557634Z * [new branch] gh/laithsakka/326/orig -> origin/gh/laithsakka/326/orig 2025-12-04T09:20:54.3557815Z * [new branch] gh/laithsakka/327/base -> origin/gh/laithsakka/327/base 2025-12-04T09:20:54.3558116Z * [new branch] gh/laithsakka/327/head -> origin/gh/laithsakka/327/head 2025-12-04T09:20:54.3558693Z * [new branch] gh/laithsakka/327/orig -> origin/gh/laithsakka/327/orig 2025-12-04T09:20:54.3559952Z * [new branch] gh/laithsakka/328/base -> origin/gh/laithsakka/328/base 2025-12-04T09:20:54.3562579Z * [new branch] gh/laithsakka/328/head -> origin/gh/laithsakka/328/head 2025-12-04T09:20:54.3562808Z * [new branch] gh/laithsakka/328/orig -> origin/gh/laithsakka/328/orig 2025-12-04T09:20:54.3563027Z * [new branch] gh/liangel/4/base -> origin/gh/liangel/4/base 2025-12-04T09:20:54.3568255Z * [new branch] gh/liangel/4/head -> origin/gh/liangel/4/head 2025-12-04T09:20:54.3568442Z * [new branch] gh/liangel/4/orig -> origin/gh/liangel/4/orig 2025-12-04T09:20:54.3568639Z * [new branch] gh/lucaskabela/1/base -> origin/gh/lucaskabela/1/base 2025-12-04T09:20:54.3568798Z * [new branch] gh/lucaskabela/1/head -> origin/gh/lucaskabela/1/head 2025-12-04T09:20:54.3568952Z * [new branch] gh/lw/4/base -> origin/gh/lw/4/base 2025-12-04T09:20:54.3569187Z * [new branch] gh/lw/4/head -> origin/gh/lw/4/head 2025-12-04T09:20:54.3570466Z * [new branch] gh/lw/4/orig -> origin/gh/lw/4/orig 2025-12-04T09:20:54.3571370Z * [new branch] gh/lw/5/base -> origin/gh/lw/5/base 2025-12-04T09:20:54.3572072Z * [new branch] gh/lw/5/head -> origin/gh/lw/5/head 2025-12-04T09:20:54.3572775Z * [new branch] gh/lw/5/orig -> origin/gh/lw/5/orig 2025-12-04T09:20:54.3573900Z * [new branch] gh/lw/6/base -> origin/gh/lw/6/base 2025-12-04T09:20:54.3574640Z * [new branch] gh/lw/6/head -> origin/gh/lw/6/head 2025-12-04T09:20:54.3575238Z * [new branch] gh/lw/6/orig -> origin/gh/lw/6/orig 2025-12-04T09:20:54.3579405Z * [new branch] gh/malfet/14/base -> origin/gh/malfet/14/base 2025-12-04T09:20:54.3579598Z * [new branch] gh/malfet/417/base -> origin/gh/malfet/417/base 2025-12-04T09:20:54.3579758Z * [new branch] gh/malfet/417/head -> origin/gh/malfet/417/head 2025-12-04T09:20:54.3579902Z * [new branch] gh/malfet/417/orig -> origin/gh/malfet/417/orig 2025-12-04T09:20:54.3580079Z * [new branch] gh/malfet/506/base -> origin/gh/malfet/506/base 2025-12-04T09:20:54.3581348Z * [new branch] gh/malfet/506/head -> origin/gh/malfet/506/head 2025-12-04T09:20:54.3581503Z * [new branch] gh/malfet/506/orig -> origin/gh/malfet/506/orig 2025-12-04T09:20:54.3582736Z * [new branch] gh/malfet/517/base -> origin/gh/malfet/517/base 2025-12-04T09:20:54.3583019Z * [new branch] gh/malfet/517/head -> origin/gh/malfet/517/head 2025-12-04T09:20:54.3584431Z * [new branch] gh/malfet/528/base -> origin/gh/malfet/528/base 2025-12-04T09:20:54.3584811Z * [new branch] gh/malfet/528/head -> origin/gh/malfet/528/head 2025-12-04T09:20:54.3586433Z * [new branch] gh/malfet/528/orig -> origin/gh/malfet/528/orig 2025-12-04T09:20:54.3586591Z * [new branch] gh/malfet/537/base -> origin/gh/malfet/537/base 2025-12-04T09:20:54.3587802Z * [new branch] gh/malfet/537/head -> origin/gh/malfet/537/head 2025-12-04T09:20:54.3588134Z * [new branch] gh/malfet/537/orig -> origin/gh/malfet/537/orig 2025-12-04T09:20:54.3589879Z * [new branch] gh/malfet/546/base -> origin/gh/malfet/546/base 2025-12-04T09:20:54.3590161Z * [new branch] gh/malfet/546/head -> origin/gh/malfet/546/head 2025-12-04T09:20:54.3590376Z * [new branch] gh/malfet/546/orig -> origin/gh/malfet/546/orig 2025-12-04T09:20:54.3591699Z * [new branch] gh/malfet/565/base -> origin/gh/malfet/565/base 2025-12-04T09:20:54.3592104Z * [new branch] gh/malfet/565/head -> origin/gh/malfet/565/head 2025-12-04T09:20:54.3593196Z * [new branch] gh/malfet/565/orig -> origin/gh/malfet/565/orig 2025-12-04T09:20:54.3594193Z * [new branch] gh/malfet/575/base -> origin/gh/malfet/575/base 2025-12-04T09:20:54.3594567Z * [new branch] gh/malfet/575/head -> origin/gh/malfet/575/head 2025-12-04T09:20:54.3595646Z * [new branch] gh/malfet/575/orig -> origin/gh/malfet/575/orig 2025-12-04T09:20:54.3598178Z * [new branch] gh/malfet/580/base -> origin/gh/malfet/580/base 2025-12-04T09:20:54.3598340Z * [new branch] gh/malfet/580/head -> origin/gh/malfet/580/head 2025-12-04T09:20:54.3598480Z * [new branch] gh/malfet/580/orig -> origin/gh/malfet/580/orig 2025-12-04T09:20:54.3599625Z * [new branch] gh/malfet/581/base -> origin/gh/malfet/581/base 2025-12-04T09:20:54.3600057Z * [new branch] gh/malfet/581/head -> origin/gh/malfet/581/head 2025-12-04T09:20:54.3601099Z * [new branch] gh/malfet/581/orig -> origin/gh/malfet/581/orig 2025-12-04T09:20:54.3602130Z * [new branch] gh/malfet/583/base -> origin/gh/malfet/583/base 2025-12-04T09:20:54.3602679Z * [new branch] gh/malfet/583/head -> origin/gh/malfet/583/head 2025-12-04T09:20:54.3603681Z * [new branch] gh/malfet/583/orig -> origin/gh/malfet/583/orig 2025-12-04T09:20:54.3604697Z * [new branch] gh/malfet/586/base -> origin/gh/malfet/586/base 2025-12-04T09:20:54.3605217Z * [new branch] gh/malfet/586/head -> origin/gh/malfet/586/head 2025-12-04T09:20:54.3605714Z * [new branch] gh/malfet/586/orig -> origin/gh/malfet/586/orig 2025-12-04T09:20:54.3608170Z * [new branch] gh/malfet/587/base -> origin/gh/malfet/587/base 2025-12-04T09:20:54.3608356Z * [new branch] gh/malfet/587/head -> origin/gh/malfet/587/head 2025-12-04T09:20:54.3608513Z * [new branch] gh/malfet/587/orig -> origin/gh/malfet/587/orig 2025-12-04T09:20:54.3609235Z * [new branch] gh/malfet/588/base -> origin/gh/malfet/588/base 2025-12-04T09:20:54.3609851Z * [new branch] gh/malfet/588/head -> origin/gh/malfet/588/head 2025-12-04T09:20:54.3611057Z * [new branch] gh/malfet/588/orig -> origin/gh/malfet/588/orig 2025-12-04T09:20:54.3611695Z * [new branch] gh/malfet/589/base -> origin/gh/malfet/589/base 2025-12-04T09:20:54.3612619Z * [new branch] gh/malfet/589/head -> origin/gh/malfet/589/head 2025-12-04T09:20:54.3613022Z * [new branch] gh/malfet/589/orig -> origin/gh/malfet/589/orig 2025-12-04T09:20:54.3616778Z * [new branch] gh/malfet/590/base -> origin/gh/malfet/590/base 2025-12-04T09:20:54.3616961Z * [new branch] gh/malfet/590/head -> origin/gh/malfet/590/head 2025-12-04T09:20:54.3617115Z * [new branch] gh/malfet/590/orig -> origin/gh/malfet/590/orig 2025-12-04T09:20:54.3617262Z * [new branch] gh/malfet/591/base -> origin/gh/malfet/591/base 2025-12-04T09:20:54.3617547Z * [new branch] gh/malfet/591/head -> origin/gh/malfet/591/head 2025-12-04T09:20:54.3619029Z * [new branch] gh/malfet/591/orig -> origin/gh/malfet/591/orig 2025-12-04T09:20:54.3619516Z * [new branch] gh/malfet/592/base -> origin/gh/malfet/592/base 2025-12-04T09:20:54.3624914Z * [new branch] gh/malfet/592/head -> origin/gh/malfet/592/head 2025-12-04T09:20:54.3629803Z * [new branch] gh/malfet/592/orig -> origin/gh/malfet/592/orig 2025-12-04T09:20:54.3634942Z * [new branch] gh/malfet/593/base -> origin/gh/malfet/593/base 2025-12-04T09:20:54.3639291Z * [new branch] gh/malfet/593/head -> origin/gh/malfet/593/head 2025-12-04T09:20:54.3639789Z * [new branch] gh/malfet/593/orig -> origin/gh/malfet/593/orig 2025-12-04T09:20:54.3639991Z * [new branch] gh/malfet/594/base -> origin/gh/malfet/594/base 2025-12-04T09:20:54.3640146Z * [new branch] gh/malfet/594/head -> origin/gh/malfet/594/head 2025-12-04T09:20:54.3640290Z * [new branch] gh/malfet/594/orig -> origin/gh/malfet/594/orig 2025-12-04T09:20:54.3640442Z * [new branch] gh/malfet/595/base -> origin/gh/malfet/595/base 2025-12-04T09:20:54.3640590Z * [new branch] gh/malfet/595/head -> origin/gh/malfet/595/head 2025-12-04T09:20:54.3640737Z * [new branch] gh/malfet/595/orig -> origin/gh/malfet/595/orig 2025-12-04T09:20:54.3640877Z * [new branch] gh/malfet/596/base -> origin/gh/malfet/596/base 2025-12-04T09:20:54.3641023Z * [new branch] gh/malfet/596/head -> origin/gh/malfet/596/head 2025-12-04T09:20:54.3641171Z * [new branch] gh/malfet/596/orig -> origin/gh/malfet/596/orig 2025-12-04T09:20:54.3641513Z * [new branch] gh/malfet/597/base -> origin/gh/malfet/597/base 2025-12-04T09:20:54.3641663Z * [new branch] gh/malfet/597/head -> origin/gh/malfet/597/head 2025-12-04T09:20:54.3642094Z * [new branch] gh/malfet/597/orig -> origin/gh/malfet/597/orig 2025-12-04T09:20:54.3642254Z * [new branch] gh/malfet/598/base -> origin/gh/malfet/598/base 2025-12-04T09:20:54.3642497Z * [new branch] gh/malfet/598/head -> origin/gh/malfet/598/head 2025-12-04T09:20:54.3642643Z * [new branch] gh/malfet/598/orig -> origin/gh/malfet/598/orig 2025-12-04T09:20:54.3642791Z * [new branch] gh/malfet/599/base -> origin/gh/malfet/599/base 2025-12-04T09:20:54.3642929Z * [new branch] gh/malfet/599/head -> origin/gh/malfet/599/head 2025-12-04T09:20:54.3643069Z * [new branch] gh/malfet/599/orig -> origin/gh/malfet/599/orig 2025-12-04T09:20:54.3643396Z * [new branch] gh/malfet/600/base -> origin/gh/malfet/600/base 2025-12-04T09:20:54.3643565Z * [new branch] gh/malfet/600/head -> origin/gh/malfet/600/head 2025-12-04T09:20:54.3643714Z * [new branch] gh/malfet/600/orig -> origin/gh/malfet/600/orig 2025-12-04T09:20:54.3645500Z * [new branch] gh/malfet/601/base -> origin/gh/malfet/601/base 2025-12-04T09:20:54.3645902Z * [new branch] gh/malfet/601/head -> origin/gh/malfet/601/head 2025-12-04T09:20:54.3646508Z * [new branch] gh/malfet/601/orig -> origin/gh/malfet/601/orig 2025-12-04T09:20:54.3650543Z * [new branch] gh/malfet/602/base -> origin/gh/malfet/602/base 2025-12-04T09:20:54.3651111Z * [new branch] gh/malfet/602/head -> origin/gh/malfet/602/head 2025-12-04T09:20:54.3651289Z * [new branch] gh/malfet/602/orig -> origin/gh/malfet/602/orig 2025-12-04T09:20:54.3651486Z * [new branch] gh/malfet/603/base -> origin/gh/malfet/603/base 2025-12-04T09:20:54.3651629Z * [new branch] gh/malfet/603/head -> origin/gh/malfet/603/head 2025-12-04T09:20:54.3652207Z * [new branch] gh/malfet/603/orig -> origin/gh/malfet/603/orig 2025-12-04T09:20:54.3653565Z * [new branch] gh/malfet/604/base -> origin/gh/malfet/604/base 2025-12-04T09:20:54.3654172Z * [new branch] gh/malfet/604/head -> origin/gh/malfet/604/head 2025-12-04T09:20:54.3654662Z * [new branch] gh/malfet/604/orig -> origin/gh/malfet/604/orig 2025-12-04T09:20:54.3654941Z * [new branch] gh/malfet/605/base -> origin/gh/malfet/605/base 2025-12-04T09:20:54.3656044Z * [new branch] gh/malfet/605/head -> origin/gh/malfet/605/head 2025-12-04T09:20:54.3657680Z * [new branch] gh/malfet/605/orig -> origin/gh/malfet/605/orig 2025-12-04T09:20:54.3657894Z * [new branch] gh/malfet/606/base -> origin/gh/malfet/606/base 2025-12-04T09:20:54.3659084Z * [new branch] gh/malfet/606/head -> origin/gh/malfet/606/head 2025-12-04T09:20:54.3659357Z * [new branch] gh/malfet/606/orig -> origin/gh/malfet/606/orig 2025-12-04T09:20:54.3660932Z * [new branch] gh/malfet/607/base -> origin/gh/malfet/607/base 2025-12-04T09:20:54.3661206Z * [new branch] gh/malfet/607/head -> origin/gh/malfet/607/head 2025-12-04T09:20:54.3661794Z * [new branch] gh/malfet/607/orig -> origin/gh/malfet/607/orig 2025-12-04T09:20:54.3665489Z * [new branch] gh/malfet/608/base -> origin/gh/malfet/608/base 2025-12-04T09:20:54.3665662Z * [new branch] gh/malfet/608/head -> origin/gh/malfet/608/head 2025-12-04T09:20:54.3665803Z * [new branch] gh/malfet/608/orig -> origin/gh/malfet/608/orig 2025-12-04T09:20:54.3666121Z * [new branch] gh/malfet/609/base -> origin/gh/malfet/609/base 2025-12-04T09:20:54.3666278Z * [new branch] gh/malfet/609/head -> origin/gh/malfet/609/head 2025-12-04T09:20:54.3670623Z * [new branch] gh/malfet/609/orig -> origin/gh/malfet/609/orig 2025-12-04T09:20:54.3671219Z * [new branch] gh/malfet/610/base -> origin/gh/malfet/610/base 2025-12-04T09:20:54.3671408Z * [new branch] gh/malfet/610/head -> origin/gh/malfet/610/head 2025-12-04T09:20:54.3671559Z * [new branch] gh/malfet/610/orig -> origin/gh/malfet/610/orig 2025-12-04T09:20:54.3671707Z * [new branch] gh/malfet/611/base -> origin/gh/malfet/611/base 2025-12-04T09:20:54.3671854Z * [new branch] gh/malfet/611/head -> origin/gh/malfet/611/head 2025-12-04T09:20:54.3672001Z * [new branch] gh/malfet/611/orig -> origin/gh/malfet/611/orig 2025-12-04T09:20:54.3674434Z * [new branch] gh/malfet/612/base -> origin/gh/malfet/612/base 2025-12-04T09:20:54.3674615Z * [new branch] gh/malfet/612/head -> origin/gh/malfet/612/head 2025-12-04T09:20:54.3674779Z * [new branch] gh/malfet/612/orig -> origin/gh/malfet/612/orig 2025-12-04T09:20:54.3674973Z * [new branch] gh/malfet/64/base -> origin/gh/malfet/64/base 2025-12-04T09:20:54.3676964Z * [new branch] gh/malfet/64/head -> origin/gh/malfet/64/head 2025-12-04T09:20:54.3677190Z * [new branch] gh/manuelcandales/11/base -> origin/gh/manuelcandales/11/base 2025-12-04T09:20:54.3677373Z * [new branch] gh/manuelcandales/11/head -> origin/gh/manuelcandales/11/head 2025-12-04T09:20:54.3677552Z * [new branch] gh/manuelcandales/11/orig -> origin/gh/manuelcandales/11/orig 2025-12-04T09:20:54.3681924Z * [new branch] gh/markkm/1/base -> origin/gh/markkm/1/base 2025-12-04T09:20:54.3682175Z * [new branch] gh/masnesral/1/base -> origin/gh/masnesral/1/base 2025-12-04T09:20:54.3682335Z * [new branch] gh/masnesral/1/head -> origin/gh/masnesral/1/head 2025-12-04T09:20:54.3682492Z * [new branch] gh/masnesral/1/orig -> origin/gh/masnesral/1/orig 2025-12-04T09:20:54.3689987Z * [new branch] gh/mhorowitz/0/base -> origin/gh/mhorowitz/0/base 2025-12-04T09:20:54.3694577Z * [new branch] gh/mhorowitz/0/head -> origin/gh/mhorowitz/0/head 2025-12-04T09:20:54.3699115Z * [new branch] gh/mhorowitz/1/base -> origin/gh/mhorowitz/1/base 2025-12-04T09:20:54.3700412Z * [new branch] gh/mhorowitz/1/head -> origin/gh/mhorowitz/1/head 2025-12-04T09:20:54.3700625Z * [new branch] gh/mhorowitz/2/base -> origin/gh/mhorowitz/2/base 2025-12-04T09:20:54.3701248Z * [new branch] gh/mhorowitz/2/head -> origin/gh/mhorowitz/2/head 2025-12-04T09:20:54.3701592Z * [new branch] gh/mhorowitz/3/base -> origin/gh/mhorowitz/3/base 2025-12-04T09:20:54.3701855Z * [new branch] gh/mhorowitz/3/head -> origin/gh/mhorowitz/3/head 2025-12-04T09:20:54.3702117Z * [new branch] gh/mhorowitz/4/base -> origin/gh/mhorowitz/4/base 2025-12-04T09:20:54.3702283Z * [new branch] gh/mhorowitz/4/head -> origin/gh/mhorowitz/4/head 2025-12-04T09:20:54.3702436Z * [new branch] gh/mhorowitz/5/base -> origin/gh/mhorowitz/5/base 2025-12-04T09:20:54.3706723Z * [new branch] gh/mhorowitz/5/head -> origin/gh/mhorowitz/5/head 2025-12-04T09:20:54.3706919Z * [new branch] gh/mhorowitz/6/base -> origin/gh/mhorowitz/6/base 2025-12-04T09:20:54.3707084Z * [new branch] gh/mhorowitz/6/head -> origin/gh/mhorowitz/6/head 2025-12-04T09:20:54.3707446Z * [new branch] gh/mikaylagawarecki/234/base -> origin/gh/mikaylagawarecki/234/base 2025-12-04T09:20:54.3707653Z * [new branch] gh/mikaylagawarecki/234/head -> origin/gh/mikaylagawarecki/234/head 2025-12-04T09:20:54.3707835Z * [new branch] gh/mikaylagawarecki/235/base -> origin/gh/mikaylagawarecki/235/base 2025-12-04T09:20:54.3708081Z * [new branch] gh/mikaylagawarecki/235/head -> origin/gh/mikaylagawarecki/235/head 2025-12-04T09:20:54.3708265Z * [new branch] gh/mikaylagawarecki/236/base -> origin/gh/mikaylagawarecki/236/base 2025-12-04T09:20:54.3708439Z * [new branch] gh/mikaylagawarecki/236/head -> origin/gh/mikaylagawarecki/236/head 2025-12-04T09:20:54.3708621Z * [new branch] gh/mikaylagawarecki/237/base -> origin/gh/mikaylagawarecki/237/base 2025-12-04T09:20:54.3708795Z * [new branch] gh/mikaylagawarecki/237/head -> origin/gh/mikaylagawarecki/237/head 2025-12-04T09:20:54.3708982Z * [new branch] gh/mikaylagawarecki/238/base -> origin/gh/mikaylagawarecki/238/base 2025-12-04T09:20:54.3709164Z * [new branch] gh/mikaylagawarecki/238/head -> origin/gh/mikaylagawarecki/238/head 2025-12-04T09:20:54.3709339Z * [new branch] gh/mikaylagawarecki/336/base -> origin/gh/mikaylagawarecki/336/base 2025-12-04T09:20:54.3709524Z * [new branch] gh/mikaylagawarecki/336/head -> origin/gh/mikaylagawarecki/336/head 2025-12-04T09:20:54.3709697Z * [new branch] gh/mikaylagawarecki/336/orig -> origin/gh/mikaylagawarecki/336/orig 2025-12-04T09:20:54.3709869Z * [new branch] gh/mikaylagawarecki/341/base -> origin/gh/mikaylagawarecki/341/base 2025-12-04T09:20:54.3710050Z * [new branch] gh/mikaylagawarecki/341/head -> origin/gh/mikaylagawarecki/341/head 2025-12-04T09:20:54.3710223Z * [new branch] gh/mikaylagawarecki/341/orig -> origin/gh/mikaylagawarecki/341/orig 2025-12-04T09:20:54.3710414Z * [new branch] gh/mikaylagawarecki/342/base -> origin/gh/mikaylagawarecki/342/base 2025-12-04T09:20:54.3710596Z * [new branch] gh/mikaylagawarecki/342/head -> origin/gh/mikaylagawarecki/342/head 2025-12-04T09:20:54.3710770Z * [new branch] gh/mikaylagawarecki/342/orig -> origin/gh/mikaylagawarecki/342/orig 2025-12-04T09:20:54.3710961Z * [new branch] gh/mikaylagawarecki/345/base -> origin/gh/mikaylagawarecki/345/base 2025-12-04T09:20:54.3711166Z * [new branch] gh/mikaylagawarecki/345/head -> origin/gh/mikaylagawarecki/345/head 2025-12-04T09:20:54.3711852Z * [new branch] gh/mikaylagawarecki/345/orig -> origin/gh/mikaylagawarecki/345/orig 2025-12-04T09:20:54.3713642Z * [new branch] gh/mikaylagawarecki/346/base -> origin/gh/mikaylagawarecki/346/base 2025-12-04T09:20:54.3713853Z * [new branch] gh/mikaylagawarecki/346/head -> origin/gh/mikaylagawarecki/346/head 2025-12-04T09:20:54.3714514Z * [new branch] gh/mikaylagawarecki/346/orig -> origin/gh/mikaylagawarecki/346/orig 2025-12-04T09:20:54.3715801Z * [new branch] gh/mikaylagawarecki/347/base -> origin/gh/mikaylagawarecki/347/base 2025-12-04T09:20:54.3716111Z * [new branch] gh/mikaylagawarecki/347/head -> origin/gh/mikaylagawarecki/347/head 2025-12-04T09:20:54.3717432Z * [new branch] gh/mikaylagawarecki/347/orig -> origin/gh/mikaylagawarecki/347/orig 2025-12-04T09:20:54.3722012Z * [new branch] gh/mikaylagawarecki/350/base -> origin/gh/mikaylagawarecki/350/base 2025-12-04T09:20:54.3722278Z * [new branch] gh/mikaylagawarecki/350/head -> origin/gh/mikaylagawarecki/350/head 2025-12-04T09:20:54.3722472Z * [new branch] gh/mikaylagawarecki/350/orig -> origin/gh/mikaylagawarecki/350/orig 2025-12-04T09:20:54.3722652Z * [new branch] gh/mikaylagawarecki/351/base -> origin/gh/mikaylagawarecki/351/base 2025-12-04T09:20:54.3722837Z * [new branch] gh/mikaylagawarecki/351/head -> origin/gh/mikaylagawarecki/351/head 2025-12-04T09:20:54.3723371Z * [new branch] gh/mikaylagawarecki/351/orig -> origin/gh/mikaylagawarecki/351/orig 2025-12-04T09:20:54.3730747Z * [new branch] gh/mikaylagawarecki/352/base -> origin/gh/mikaylagawarecki/352/base 2025-12-04T09:20:54.3735499Z * [new branch] gh/mikaylagawarecki/352/head -> origin/gh/mikaylagawarecki/352/head 2025-12-04T09:20:54.3740128Z * [new branch] gh/mikaylagawarecki/352/orig -> origin/gh/mikaylagawarecki/352/orig 2025-12-04T09:20:54.3744829Z * [new branch] gh/mikaylagawarecki/353/base -> origin/gh/mikaylagawarecki/353/base 2025-12-04T09:20:54.3749218Z * [new branch] gh/mikaylagawarecki/353/head -> origin/gh/mikaylagawarecki/353/head 2025-12-04T09:20:54.3754344Z * [new branch] gh/mikaylagawarecki/353/orig -> origin/gh/mikaylagawarecki/353/orig 2025-12-04T09:20:54.3756764Z * [new branch] gh/mikaylagawarecki/354/base -> origin/gh/mikaylagawarecki/354/base 2025-12-04T09:20:54.3756989Z * [new branch] gh/mikaylagawarecki/354/head -> origin/gh/mikaylagawarecki/354/head 2025-12-04T09:20:54.3757216Z * [new branch] gh/mikaylagawarecki/354/orig -> origin/gh/mikaylagawarecki/354/orig 2025-12-04T09:20:54.3757394Z * [new branch] gh/mikaylagawarecki/356/base -> origin/gh/mikaylagawarecki/356/base 2025-12-04T09:20:54.3757601Z * [new branch] gh/mikaylagawarecki/356/head -> origin/gh/mikaylagawarecki/356/head 2025-12-04T09:20:54.3757776Z * [new branch] gh/mikaylagawarecki/356/orig -> origin/gh/mikaylagawarecki/356/orig 2025-12-04T09:20:54.3757953Z * [new branch] gh/mikaylagawarecki/357/base -> origin/gh/mikaylagawarecki/357/base 2025-12-04T09:20:54.3758143Z * [new branch] gh/mikaylagawarecki/357/head -> origin/gh/mikaylagawarecki/357/head 2025-12-04T09:20:54.3758321Z * [new branch] gh/mikaylagawarecki/357/orig -> origin/gh/mikaylagawarecki/357/orig 2025-12-04T09:20:54.3758511Z * [new branch] gh/mikaylagawarecki/359/base -> origin/gh/mikaylagawarecki/359/base 2025-12-04T09:20:54.3758685Z * [new branch] gh/mikaylagawarecki/359/head -> origin/gh/mikaylagawarecki/359/head 2025-12-04T09:20:54.3758861Z * [new branch] gh/mikaylagawarecki/359/orig -> origin/gh/mikaylagawarecki/359/orig 2025-12-04T09:20:54.3759048Z * [new branch] gh/mikaylagawarecki/360/base -> origin/gh/mikaylagawarecki/360/base 2025-12-04T09:20:54.3759220Z * [new branch] gh/mikaylagawarecki/360/head -> origin/gh/mikaylagawarecki/360/head 2025-12-04T09:20:54.3759403Z * [new branch] gh/mikaylagawarecki/360/orig -> origin/gh/mikaylagawarecki/360/orig 2025-12-04T09:20:54.3759578Z * [new branch] gh/mikaylagawarecki/361/base -> origin/gh/mikaylagawarecki/361/base 2025-12-04T09:20:54.3759752Z * [new branch] gh/mikaylagawarecki/361/head -> origin/gh/mikaylagawarecki/361/head 2025-12-04T09:20:54.3759937Z * [new branch] gh/mikaylagawarecki/361/orig -> origin/gh/mikaylagawarecki/361/orig 2025-12-04T09:20:54.3760111Z * [new branch] gh/mikaylagawarecki/362/base -> origin/gh/mikaylagawarecki/362/base 2025-12-04T09:20:54.3760292Z * [new branch] gh/mikaylagawarecki/362/head -> origin/gh/mikaylagawarecki/362/head 2025-12-04T09:20:54.3760468Z * [new branch] gh/mikaylagawarecki/362/orig -> origin/gh/mikaylagawarecki/362/orig 2025-12-04T09:20:54.3760642Z * [new branch] gh/mikaylagawarecki/363/base -> origin/gh/mikaylagawarecki/363/base 2025-12-04T09:20:54.3760825Z * [new branch] gh/mikaylagawarecki/363/head -> origin/gh/mikaylagawarecki/363/head 2025-12-04T09:20:54.3761002Z * [new branch] gh/mikaylagawarecki/363/orig -> origin/gh/mikaylagawarecki/363/orig 2025-12-04T09:20:54.3761184Z * [new branch] gh/mikaylagawarecki/364/base -> origin/gh/mikaylagawarecki/364/base 2025-12-04T09:20:54.3761836Z * [new branch] gh/mikaylagawarecki/364/head -> origin/gh/mikaylagawarecki/364/head 2025-12-04T09:20:54.3762032Z * [new branch] gh/mikaylagawarecki/364/orig -> origin/gh/mikaylagawarecki/364/orig 2025-12-04T09:20:54.3762222Z * [new branch] gh/mikaylagawarecki/365/base -> origin/gh/mikaylagawarecki/365/base 2025-12-04T09:20:54.3762457Z * [new branch] gh/mikaylagawarecki/365/head -> origin/gh/mikaylagawarecki/365/head 2025-12-04T09:20:54.3762645Z * [new branch] gh/mikaylagawarecki/365/orig -> origin/gh/mikaylagawarecki/365/orig 2025-12-04T09:20:54.3762821Z * [new branch] gh/mikaylagawarecki/366/base -> origin/gh/mikaylagawarecki/366/base 2025-12-04T09:20:54.3763004Z * [new branch] gh/mikaylagawarecki/366/head -> origin/gh/mikaylagawarecki/366/head 2025-12-04T09:20:54.3763197Z * [new branch] gh/mikaylagawarecki/366/orig -> origin/gh/mikaylagawarecki/366/orig 2025-12-04T09:20:54.3763375Z * [new branch] gh/mikaylagawarecki/367/base -> origin/gh/mikaylagawarecki/367/base 2025-12-04T09:20:54.3763561Z * [new branch] gh/mikaylagawarecki/367/head -> origin/gh/mikaylagawarecki/367/head 2025-12-04T09:20:54.3768281Z * [new branch] gh/mikaylagawarecki/367/orig -> origin/gh/mikaylagawarecki/367/orig 2025-12-04T09:20:54.3770322Z * [new branch] gh/mikaylagawarecki/368/base -> origin/gh/mikaylagawarecki/368/base 2025-12-04T09:20:54.3770531Z * [new branch] gh/mikaylagawarecki/368/head -> origin/gh/mikaylagawarecki/368/head 2025-12-04T09:20:54.3770721Z * [new branch] gh/mikaylagawarecki/368/orig -> origin/gh/mikaylagawarecki/368/orig 2025-12-04T09:20:54.3770917Z * [new branch] gh/mikaylagawarecki/369/base -> origin/gh/mikaylagawarecki/369/base 2025-12-04T09:20:54.3771107Z * [new branch] gh/mikaylagawarecki/369/head -> origin/gh/mikaylagawarecki/369/head 2025-12-04T09:20:54.3774225Z * [new branch] gh/mikaylagawarecki/369/orig -> origin/gh/mikaylagawarecki/369/orig 2025-12-04T09:20:54.3774473Z * [new branch] gh/mikaylagawarecki/370/base -> origin/gh/mikaylagawarecki/370/base 2025-12-04T09:20:54.3774655Z * [new branch] gh/mikaylagawarecki/370/head -> origin/gh/mikaylagawarecki/370/head 2025-12-04T09:20:54.3774850Z * [new branch] gh/mikaylagawarecki/370/orig -> origin/gh/mikaylagawarecki/370/orig 2025-12-04T09:20:54.3775028Z * [new branch] gh/mikaylagawarecki/371/base -> origin/gh/mikaylagawarecki/371/base 2025-12-04T09:20:54.3775204Z * [new branch] gh/mikaylagawarecki/371/head -> origin/gh/mikaylagawarecki/371/head 2025-12-04T09:20:54.3776457Z * [new branch] gh/mikaylagawarecki/371/orig -> origin/gh/mikaylagawarecki/371/orig 2025-12-04T09:20:54.3776643Z * [new branch] gh/mikaylagawarecki/372/base -> origin/gh/mikaylagawarecki/372/base 2025-12-04T09:20:54.3776825Z * [new branch] gh/mikaylagawarecki/372/head -> origin/gh/mikaylagawarecki/372/head 2025-12-04T09:20:54.3777008Z * [new branch] gh/mikaylagawarecki/372/orig -> origin/gh/mikaylagawarecki/372/orig 2025-12-04T09:20:54.3777186Z * [new branch] gh/mikaylagawarecki/373/base -> origin/gh/mikaylagawarecki/373/base 2025-12-04T09:20:54.3777368Z * [new branch] gh/mikaylagawarecki/373/head -> origin/gh/mikaylagawarecki/373/head 2025-12-04T09:20:54.3783421Z * [new branch] gh/mikaylagawarecki/373/orig -> origin/gh/mikaylagawarecki/373/orig 2025-12-04T09:20:54.3783661Z * [new branch] gh/mikaylagawarecki/374/base -> origin/gh/mikaylagawarecki/374/base 2025-12-04T09:20:54.3783851Z * [new branch] gh/mikaylagawarecki/374/head -> origin/gh/mikaylagawarecki/374/head 2025-12-04T09:20:54.3784028Z * [new branch] gh/mikaylagawarecki/374/orig -> origin/gh/mikaylagawarecki/374/orig 2025-12-04T09:20:54.3784206Z * [new branch] gh/mikaylagawarecki/375/base -> origin/gh/mikaylagawarecki/375/base 2025-12-04T09:20:54.3784573Z * [new branch] gh/mikaylagawarecki/375/head -> origin/gh/mikaylagawarecki/375/head 2025-12-04T09:20:54.3784753Z * [new branch] gh/mikaylagawarecki/375/orig -> origin/gh/mikaylagawarecki/375/orig 2025-12-04T09:20:54.3787355Z * [new branch] gh/mikaylagawarecki/376/base -> origin/gh/mikaylagawarecki/376/base 2025-12-04T09:20:54.3787716Z * [new branch] gh/mikaylagawarecki/376/head -> origin/gh/mikaylagawarecki/376/head 2025-12-04T09:20:54.3787946Z * [new branch] gh/mikaylagawarecki/376/orig -> origin/gh/mikaylagawarecki/376/orig 2025-12-04T09:20:54.3788135Z * [new branch] gh/mikaylagawarecki/377/base -> origin/gh/mikaylagawarecki/377/base 2025-12-04T09:20:54.3788374Z * [new branch] gh/mikaylagawarecki/377/head -> origin/gh/mikaylagawarecki/377/head 2025-12-04T09:20:54.3788567Z * [new branch] gh/mikaylagawarecki/377/orig -> origin/gh/mikaylagawarecki/377/orig 2025-12-04T09:20:54.3792775Z * [new branch] gh/mikaylagawarecki/378/base -> origin/gh/mikaylagawarecki/378/base 2025-12-04T09:20:54.3793022Z * [new branch] gh/mikaylagawarecki/378/head -> origin/gh/mikaylagawarecki/378/head 2025-12-04T09:20:54.3793209Z * [new branch] gh/mikaylagawarecki/378/orig -> origin/gh/mikaylagawarecki/378/orig 2025-12-04T09:20:54.3793521Z * [new branch] gh/mikaylagawarecki/379/base -> origin/gh/mikaylagawarecki/379/base 2025-12-04T09:20:54.3797826Z * [new branch] gh/mikaylagawarecki/379/head -> origin/gh/mikaylagawarecki/379/head 2025-12-04T09:20:54.3798219Z * [new branch] gh/mikaylagawarecki/379/orig -> origin/gh/mikaylagawarecki/379/orig 2025-12-04T09:20:54.3798503Z * [new branch] gh/mikaylagawarecki/380/base -> origin/gh/mikaylagawarecki/380/base 2025-12-04T09:20:54.3798720Z * [new branch] gh/mikaylagawarecki/380/head -> origin/gh/mikaylagawarecki/380/head 2025-12-04T09:20:54.3798913Z * [new branch] gh/mikaylagawarecki/380/orig -> origin/gh/mikaylagawarecki/380/orig 2025-12-04T09:20:54.3799227Z * [new branch] gh/mikaylagawarecki/381/base -> origin/gh/mikaylagawarecki/381/base 2025-12-04T09:20:54.3799451Z * [new branch] gh/mikaylagawarecki/381/head -> origin/gh/mikaylagawarecki/381/head 2025-12-04T09:20:54.3800156Z * [new branch] gh/mikaylagawarecki/381/orig -> origin/gh/mikaylagawarecki/381/orig 2025-12-04T09:20:54.3800404Z * [new branch] gh/mikaylagawarecki/382/base -> origin/gh/mikaylagawarecki/382/base 2025-12-04T09:20:54.3800585Z * [new branch] gh/mikaylagawarecki/382/head -> origin/gh/mikaylagawarecki/382/head 2025-12-04T09:20:54.3800761Z * [new branch] gh/mikaylagawarecki/382/orig -> origin/gh/mikaylagawarecki/382/orig 2025-12-04T09:20:54.3800945Z * [new branch] gh/mikaylagawarecki/383/base -> origin/gh/mikaylagawarecki/383/base 2025-12-04T09:20:54.3801156Z * [new branch] gh/mikaylagawarecki/383/head -> origin/gh/mikaylagawarecki/383/head 2025-12-04T09:20:54.3801509Z * [new branch] gh/mikaylagawarecki/383/orig -> origin/gh/mikaylagawarecki/383/orig 2025-12-04T09:20:54.3807618Z * [new branch] gh/mikaylagawarecki/384/base -> origin/gh/mikaylagawarecki/384/base 2025-12-04T09:20:54.3807935Z * [new branch] gh/mikaylagawarecki/384/head -> origin/gh/mikaylagawarecki/384/head 2025-12-04T09:20:54.3808124Z * [new branch] gh/mikaylagawarecki/384/orig -> origin/gh/mikaylagawarecki/384/orig 2025-12-04T09:20:54.3808309Z * [new branch] gh/mikaylagawarecki/385/base -> origin/gh/mikaylagawarecki/385/base 2025-12-04T09:20:54.3808488Z * [new branch] gh/mikaylagawarecki/385/head -> origin/gh/mikaylagawarecki/385/head 2025-12-04T09:20:54.3808662Z * [new branch] gh/mikaylagawarecki/385/orig -> origin/gh/mikaylagawarecki/385/orig 2025-12-04T09:20:54.3808998Z * [new branch] gh/mikaylagawarecki/386/base -> origin/gh/mikaylagawarecki/386/base 2025-12-04T09:20:54.3815810Z * [new branch] gh/mikaylagawarecki/386/head -> origin/gh/mikaylagawarecki/386/head 2025-12-04T09:20:54.3816042Z * [new branch] gh/mikaylagawarecki/386/orig -> origin/gh/mikaylagawarecki/386/orig 2025-12-04T09:20:54.3816372Z * [new branch] gh/mikaylagawarecki/387/base -> origin/gh/mikaylagawarecki/387/base 2025-12-04T09:20:54.3816700Z * [new branch] gh/mikaylagawarecki/387/head -> origin/gh/mikaylagawarecki/387/head 2025-12-04T09:20:54.3816896Z * [new branch] gh/mikaylagawarecki/387/orig -> origin/gh/mikaylagawarecki/387/orig 2025-12-04T09:20:54.3817071Z * [new branch] gh/mikaylagawarecki/388/base -> origin/gh/mikaylagawarecki/388/base 2025-12-04T09:20:54.3817388Z * [new branch] gh/mikaylagawarecki/388/head -> origin/gh/mikaylagawarecki/388/head 2025-12-04T09:20:54.3818076Z * [new branch] gh/mikaylagawarecki/388/orig -> origin/gh/mikaylagawarecki/388/orig 2025-12-04T09:20:54.3818329Z * [new branch] gh/mikaylagawarecki/389/base -> origin/gh/mikaylagawarecki/389/base 2025-12-04T09:20:54.3818513Z * [new branch] gh/mikaylagawarecki/389/head -> origin/gh/mikaylagawarecki/389/head 2025-12-04T09:20:54.3818704Z * [new branch] gh/mikaylagawarecki/389/orig -> origin/gh/mikaylagawarecki/389/orig 2025-12-04T09:20:54.3818889Z * [new branch] gh/mikaylagawarecki/390/base -> origin/gh/mikaylagawarecki/390/base 2025-12-04T09:20:54.3819062Z * [new branch] gh/mikaylagawarecki/390/head -> origin/gh/mikaylagawarecki/390/head 2025-12-04T09:20:54.3819454Z * [new branch] gh/mikaylagawarecki/390/orig -> origin/gh/mikaylagawarecki/390/orig 2025-12-04T09:20:54.3820073Z * [new branch] gh/mikaylagawarecki/391/base -> origin/gh/mikaylagawarecki/391/base 2025-12-04T09:20:54.3820295Z * [new branch] gh/mikaylagawarecki/391/head -> origin/gh/mikaylagawarecki/391/head 2025-12-04T09:20:54.3820778Z * [new branch] gh/mikaylagawarecki/391/orig -> origin/gh/mikaylagawarecki/391/orig 2025-12-04T09:20:54.3822142Z * [new branch] gh/mikaylagawarecki/392/base -> origin/gh/mikaylagawarecki/392/base 2025-12-04T09:20:54.3822771Z * [new branch] gh/mikaylagawarecki/392/head -> origin/gh/mikaylagawarecki/392/head 2025-12-04T09:20:54.3823893Z * [new branch] gh/mikaylagawarecki/392/orig -> origin/gh/mikaylagawarecki/392/orig 2025-12-04T09:20:54.3825328Z * [new branch] gh/mlazos/41/base -> origin/gh/mlazos/41/base 2025-12-04T09:20:54.3825792Z * [new branch] gh/mlazos/41/head -> origin/gh/mlazos/41/head 2025-12-04T09:20:54.3826871Z * [new branch] gh/mlazos/41/orig -> origin/gh/mlazos/41/orig 2025-12-04T09:20:54.3828005Z * [new branch] gh/mlazos/42/base -> origin/gh/mlazos/42/base 2025-12-04T09:20:54.3828256Z * [new branch] gh/mlazos/42/head -> origin/gh/mlazos/42/head 2025-12-04T09:20:54.3829363Z * [new branch] gh/mlazos/42/orig -> origin/gh/mlazos/42/orig 2025-12-04T09:20:54.3829993Z * [new branch] gh/mlazos/43/base -> origin/gh/mlazos/43/base 2025-12-04T09:20:54.3831249Z * [new branch] gh/mlazos/43/head -> origin/gh/mlazos/43/head 2025-12-04T09:20:54.3831488Z * [new branch] gh/mlazos/43/orig -> origin/gh/mlazos/43/orig 2025-12-04T09:20:54.3832732Z * [new branch] gh/mlazos/44/base -> origin/gh/mlazos/44/base 2025-12-04T09:20:54.3833020Z * [new branch] gh/mlazos/44/head -> origin/gh/mlazos/44/head 2025-12-04T09:20:54.3834203Z * [new branch] gh/mlazos/44/orig -> origin/gh/mlazos/44/orig 2025-12-04T09:20:54.3835571Z * [new branch] gh/mlazos/47/base -> origin/gh/mlazos/47/base 2025-12-04T09:20:54.3836037Z * [new branch] gh/mlazos/47/head -> origin/gh/mlazos/47/head 2025-12-04T09:20:54.3837229Z * [new branch] gh/mlazos/47/orig -> origin/gh/mlazos/47/orig 2025-12-04T09:20:54.3837664Z * [new branch] gh/mlazos/48/base -> origin/gh/mlazos/48/base 2025-12-04T09:20:54.3838855Z * [new branch] gh/mlazos/48/head -> origin/gh/mlazos/48/head 2025-12-04T09:20:54.3839003Z * [new branch] gh/mlazos/48/orig -> origin/gh/mlazos/48/orig 2025-12-04T09:20:54.3840400Z * [new branch] gh/mlazos/49/base -> origin/gh/mlazos/49/base 2025-12-04T09:20:54.3840907Z * [new branch] gh/mlazos/49/head -> origin/gh/mlazos/49/head 2025-12-04T09:20:54.3842068Z * [new branch] gh/mlazos/49/orig -> origin/gh/mlazos/49/orig 2025-12-04T09:20:54.3842670Z * [new branch] gh/mlazos/50/base -> origin/gh/mlazos/50/base 2025-12-04T09:20:54.3846375Z * [new branch] gh/mlazos/50/head -> origin/gh/mlazos/50/head 2025-12-04T09:20:54.3846558Z * [new branch] gh/mlazos/50/orig -> origin/gh/mlazos/50/orig 2025-12-04T09:20:54.3846716Z * [new branch] gh/mlazos/51/base -> origin/gh/mlazos/51/base 2025-12-04T09:20:54.3846891Z * [new branch] gh/mlazos/51/head -> origin/gh/mlazos/51/head 2025-12-04T09:20:54.3847032Z * [new branch] gh/mlazos/51/orig -> origin/gh/mlazos/51/orig 2025-12-04T09:20:54.3847573Z * [new branch] gh/mlazos/52/base -> origin/gh/mlazos/52/base 2025-12-04T09:20:54.3848207Z * [new branch] gh/mlazos/52/head -> origin/gh/mlazos/52/head 2025-12-04T09:20:54.3849084Z * [new branch] gh/mlazos/52/orig -> origin/gh/mlazos/52/orig 2025-12-04T09:20:54.3850442Z * [new branch] gh/mlazos/53/base -> origin/gh/mlazos/53/base 2025-12-04T09:20:54.3850624Z * [new branch] gh/mlazos/53/head -> origin/gh/mlazos/53/head 2025-12-04T09:20:54.3851642Z * [new branch] gh/mlazos/53/orig -> origin/gh/mlazos/53/orig 2025-12-04T09:20:54.3852559Z * [new branch] gh/mlazos/54/base -> origin/gh/mlazos/54/base 2025-12-04T09:20:54.3852998Z * [new branch] gh/mlazos/54/head -> origin/gh/mlazos/54/head 2025-12-04T09:20:54.3854029Z * [new branch] gh/mlazos/54/orig -> origin/gh/mlazos/54/orig 2025-12-04T09:20:54.3855331Z * [new branch] gh/mlazos/55/base -> origin/gh/mlazos/55/base 2025-12-04T09:20:54.3855972Z * [new branch] gh/mlazos/55/head -> origin/gh/mlazos/55/head 2025-12-04T09:20:54.3856631Z * [new branch] gh/mlazos/55/orig -> origin/gh/mlazos/55/orig 2025-12-04T09:20:54.3857924Z * [new branch] gh/mlazos/56/base -> origin/gh/mlazos/56/base 2025-12-04T09:20:54.3858224Z * [new branch] gh/mlazos/56/head -> origin/gh/mlazos/56/head 2025-12-04T09:20:54.3859269Z * [new branch] gh/mlazos/56/orig -> origin/gh/mlazos/56/orig 2025-12-04T09:20:54.3862953Z * [new branch] gh/mlazos/57/base -> origin/gh/mlazos/57/base 2025-12-04T09:20:54.3863162Z * [new branch] gh/mlazos/57/head -> origin/gh/mlazos/57/head 2025-12-04T09:20:54.3863326Z * [new branch] gh/mlazos/57/orig -> origin/gh/mlazos/57/orig 2025-12-04T09:20:54.3863482Z * [new branch] gh/mlazos/58/base -> origin/gh/mlazos/58/base 2025-12-04T09:20:54.3863927Z * [new branch] gh/mlazos/58/head -> origin/gh/mlazos/58/head 2025-12-04T09:20:54.3864896Z * [new branch] gh/mlazos/58/orig -> origin/gh/mlazos/58/orig 2025-12-04T09:20:54.3868812Z * [new branch] gh/mlazos/59/base -> origin/gh/mlazos/59/base 2025-12-04T09:20:54.3869195Z * [new branch] gh/mlazos/59/head -> origin/gh/mlazos/59/head 2025-12-04T09:20:54.3869366Z * [new branch] gh/mlazos/59/orig -> origin/gh/mlazos/59/orig 2025-12-04T09:20:54.3869528Z * [new branch] gh/mlazos/60/base -> origin/gh/mlazos/60/base 2025-12-04T09:20:54.3869736Z * [new branch] gh/mlazos/60/head -> origin/gh/mlazos/60/head 2025-12-04T09:20:54.3869890Z * [new branch] gh/mlazos/60/orig -> origin/gh/mlazos/60/orig 2025-12-04T09:20:54.3874683Z * [new branch] gh/mlazos/61/base -> origin/gh/mlazos/61/base 2025-12-04T09:20:54.3874864Z * [new branch] gh/mlazos/61/head -> origin/gh/mlazos/61/head 2025-12-04T09:20:54.3875021Z * [new branch] gh/mlazos/61/orig -> origin/gh/mlazos/61/orig 2025-12-04T09:20:54.3875175Z * [new branch] gh/mlazos/62/base -> origin/gh/mlazos/62/base 2025-12-04T09:20:54.3875340Z * [new branch] gh/mlazos/62/head -> origin/gh/mlazos/62/head 2025-12-04T09:20:54.3875499Z * [new branch] gh/mlazos/62/orig -> origin/gh/mlazos/62/orig 2025-12-04T09:20:54.3877633Z * [new branch] gh/mlazos/63/base -> origin/gh/mlazos/63/base 2025-12-04T09:20:54.3877834Z * [new branch] gh/mlazos/63/head -> origin/gh/mlazos/63/head 2025-12-04T09:20:54.3877984Z * [new branch] gh/mlazos/63/orig -> origin/gh/mlazos/63/orig 2025-12-04T09:20:54.3879162Z * [new branch] gh/mlazos/64/base -> origin/gh/mlazos/64/base 2025-12-04T09:20:54.3879490Z * [new branch] gh/mlazos/64/head -> origin/gh/mlazos/64/head 2025-12-04T09:20:54.3880645Z * [new branch] gh/mlazos/64/orig -> origin/gh/mlazos/64/orig 2025-12-04T09:20:54.3881504Z * [new branch] gh/mlazos/65/base -> origin/gh/mlazos/65/base 2025-12-04T09:20:54.3882180Z * [new branch] gh/mlazos/65/head -> origin/gh/mlazos/65/head 2025-12-04T09:20:54.3886872Z * [new branch] gh/mlazos/65/orig -> origin/gh/mlazos/65/orig 2025-12-04T09:20:54.3887143Z * [new branch] gh/mlazos/66/base -> origin/gh/mlazos/66/base 2025-12-04T09:20:54.3887320Z * [new branch] gh/mlazos/66/head -> origin/gh/mlazos/66/head 2025-12-04T09:20:54.3887564Z * [new branch] gh/mlazos/66/orig -> origin/gh/mlazos/66/orig 2025-12-04T09:20:54.3887818Z * [new branch] gh/mlazos/67/base -> origin/gh/mlazos/67/base 2025-12-04T09:20:54.3887975Z * [new branch] gh/mlazos/67/head -> origin/gh/mlazos/67/head 2025-12-04T09:20:54.3888113Z * [new branch] gh/mlazos/67/orig -> origin/gh/mlazos/67/orig 2025-12-04T09:20:54.3890459Z * [new branch] gh/mlazos/68/base -> origin/gh/mlazos/68/base 2025-12-04T09:20:54.3896185Z * [new branch] gh/mlazos/68/head -> origin/gh/mlazos/68/head 2025-12-04T09:20:54.3898401Z * [new branch] gh/mlazos/68/orig -> origin/gh/mlazos/68/orig 2025-12-04T09:20:54.3898590Z * [new branch] gh/mlazos/69/base -> origin/gh/mlazos/69/base 2025-12-04T09:20:54.3898760Z * [new branch] gh/mlazos/69/head -> origin/gh/mlazos/69/head 2025-12-04T09:20:54.3898910Z * [new branch] gh/mlazos/69/orig -> origin/gh/mlazos/69/orig 2025-12-04T09:20:54.3899054Z * [new branch] gh/mlazos/70/base -> origin/gh/mlazos/70/base 2025-12-04T09:20:54.3899201Z * [new branch] gh/mlazos/70/head -> origin/gh/mlazos/70/head 2025-12-04T09:20:54.3899342Z * [new branch] gh/mlazos/70/orig -> origin/gh/mlazos/70/orig 2025-12-04T09:20:54.3899488Z * [new branch] gh/mlazos/71/base -> origin/gh/mlazos/71/base 2025-12-04T09:20:54.3899801Z * [new branch] gh/mlazos/71/head -> origin/gh/mlazos/71/head 2025-12-04T09:20:54.3899939Z * [new branch] gh/mlazos/71/orig -> origin/gh/mlazos/71/orig 2025-12-04T09:20:54.3904510Z * [new branch] gh/mlazos/72/base -> origin/gh/mlazos/72/base 2025-12-04T09:20:54.3909194Z * [new branch] gh/mlazos/72/head -> origin/gh/mlazos/72/head 2025-12-04T09:20:54.3913659Z * [new branch] gh/mlazos/72/orig -> origin/gh/mlazos/72/orig 2025-12-04T09:20:54.3913866Z * [new branch] gh/mlazos/73/base -> origin/gh/mlazos/73/base 2025-12-04T09:20:54.3914454Z * [new branch] gh/mlazos/73/head -> origin/gh/mlazos/73/head 2025-12-04T09:20:54.3914625Z * [new branch] gh/mlazos/73/orig -> origin/gh/mlazos/73/orig 2025-12-04T09:20:54.3914780Z * [new branch] gh/mrmiywj/1/base -> origin/gh/mrmiywj/1/base 2025-12-04T09:20:54.3914946Z * [new branch] gh/mrmiywj/1/head -> origin/gh/mrmiywj/1/head 2025-12-04T09:20:54.3915110Z * [new branch] gh/muchulee8/73/base -> origin/gh/muchulee8/73/base 2025-12-04T09:20:54.3915266Z * [new branch] gh/muchulee8/73/head -> origin/gh/muchulee8/73/head 2025-12-04T09:20:54.3915416Z * [new branch] gh/muchulee8/73/orig -> origin/gh/muchulee8/73/orig 2025-12-04T09:20:54.3915602Z * [new branch] gh/naveenthangudu/1/base -> origin/gh/naveenthangudu/1/base 2025-12-04T09:20:54.3915772Z * [new branch] gh/naveenthangudu/1/head -> origin/gh/naveenthangudu/1/head 2025-12-04T09:20:54.3915932Z * [new branch] gh/naveenthangudu/1/orig -> origin/gh/naveenthangudu/1/orig 2025-12-04T09:20:54.3916096Z * [new branch] gh/naveenthangudu/2/base -> origin/gh/naveenthangudu/2/base 2025-12-04T09:20:54.3916264Z * [new branch] gh/naveenthangudu/2/head -> origin/gh/naveenthangudu/2/head 2025-12-04T09:20:54.3916421Z * [new branch] gh/naveenthangudu/2/orig -> origin/gh/naveenthangudu/2/orig 2025-12-04T09:20:54.3916589Z * [new branch] gh/naveenthangudu/3/base -> origin/gh/naveenthangudu/3/base 2025-12-04T09:20:54.3916742Z * [new branch] gh/naveenthangudu/3/head -> origin/gh/naveenthangudu/3/head 2025-12-04T09:20:54.3916909Z * [new branch] gh/naveenthangudu/3/orig -> origin/gh/naveenthangudu/3/orig 2025-12-04T09:20:54.3917201Z * [new branch] gh/naveenthangudu/4/base -> origin/gh/naveenthangudu/4/base 2025-12-04T09:20:54.3917782Z * [new branch] gh/naveenthangudu/4/head -> origin/gh/naveenthangudu/4/head 2025-12-04T09:20:54.3919042Z * [new branch] gh/naveenthangudu/4/orig -> origin/gh/naveenthangudu/4/orig 2025-12-04T09:20:54.3919987Z * [new branch] gh/naveenthangudu/5/base -> origin/gh/naveenthangudu/5/base 2025-12-04T09:20:54.3920427Z * [new branch] gh/naveenthangudu/5/head -> origin/gh/naveenthangudu/5/head 2025-12-04T09:20:54.3921608Z * [new branch] gh/naveenthangudu/5/orig -> origin/gh/naveenthangudu/5/orig 2025-12-04T09:20:54.3923268Z * [new branch] gh/naveenthangudu/6/base -> origin/gh/naveenthangudu/6/base 2025-12-04T09:20:54.3923558Z * [new branch] gh/naveenthangudu/6/head -> origin/gh/naveenthangudu/6/head 2025-12-04T09:20:54.3923737Z * [new branch] gh/naveenthangudu/6/orig -> origin/gh/naveenthangudu/6/orig 2025-12-04T09:20:54.3926156Z * [new branch] gh/naveenthangudu/7/base -> origin/gh/naveenthangudu/7/base 2025-12-04T09:20:54.3927421Z * [new branch] gh/naveenthangudu/7/head -> origin/gh/naveenthangudu/7/head 2025-12-04T09:20:54.3927611Z * [new branch] gh/naveenthangudu/7/orig -> origin/gh/naveenthangudu/7/orig 2025-12-04T09:20:54.3927781Z * [new branch] gh/naveenthangudu/8/base -> origin/gh/naveenthangudu/8/base 2025-12-04T09:20:54.3929383Z * [new branch] gh/naveenthangudu/8/head -> origin/gh/naveenthangudu/8/head 2025-12-04T09:20:54.3929558Z * [new branch] gh/naveenthangudu/8/orig -> origin/gh/naveenthangudu/8/orig 2025-12-04T09:20:54.3938533Z * [new branch] gh/naveenthangudu/9/base -> origin/gh/naveenthangudu/9/base 2025-12-04T09:20:54.3938753Z * [new branch] gh/naveenthangudu/9/head -> origin/gh/naveenthangudu/9/head 2025-12-04T09:20:54.3938919Z * [new branch] gh/naveenthangudu/9/orig -> origin/gh/naveenthangudu/9/orig 2025-12-04T09:20:54.3939092Z * [new branch] gh/nikitaved/1/base -> origin/gh/nikitaved/1/base 2025-12-04T09:20:54.3939246Z * [new branch] gh/nikitaved/1/head -> origin/gh/nikitaved/1/head 2025-12-04T09:20:54.3939406Z * [new branch] gh/nikitaved/1/orig -> origin/gh/nikitaved/1/orig 2025-12-04T09:20:54.3939573Z * [new branch] gh/nikitaved/10/base -> origin/gh/nikitaved/10/base 2025-12-04T09:20:54.3939727Z * [new branch] gh/nikitaved/10/head -> origin/gh/nikitaved/10/head 2025-12-04T09:20:54.3939887Z * [new branch] gh/nikitaved/10/orig -> origin/gh/nikitaved/10/orig 2025-12-04T09:20:54.3940039Z * [new branch] gh/nikitaved/11/base -> origin/gh/nikitaved/11/base 2025-12-04T09:20:54.3940191Z * [new branch] gh/nikitaved/11/head -> origin/gh/nikitaved/11/head 2025-12-04T09:20:54.3940515Z * [new branch] gh/nikitaved/11/orig -> origin/gh/nikitaved/11/orig 2025-12-04T09:20:54.3946541Z * [new branch] gh/nikitaved/12/base -> origin/gh/nikitaved/12/base 2025-12-04T09:20:54.3946732Z * [new branch] gh/nikitaved/12/head -> origin/gh/nikitaved/12/head 2025-12-04T09:20:54.3947102Z * [new branch] gh/nikitaved/12/orig -> origin/gh/nikitaved/12/orig 2025-12-04T09:20:54.3947267Z * [new branch] gh/nikitaved/13/base -> origin/gh/nikitaved/13/base 2025-12-04T09:20:54.3947423Z * [new branch] gh/nikitaved/13/head -> origin/gh/nikitaved/13/head 2025-12-04T09:20:54.3947568Z * [new branch] gh/nikitaved/13/orig -> origin/gh/nikitaved/13/orig 2025-12-04T09:20:54.3947721Z * [new branch] gh/nikitaved/14/base -> origin/gh/nikitaved/14/base 2025-12-04T09:20:54.3947879Z * [new branch] gh/nikitaved/14/head -> origin/gh/nikitaved/14/head 2025-12-04T09:20:54.3951014Z * [new branch] gh/nikitaved/14/orig -> origin/gh/nikitaved/14/orig 2025-12-04T09:20:54.3951180Z * [new branch] gh/nikitaved/15/base -> origin/gh/nikitaved/15/base 2025-12-04T09:20:54.3951327Z * [new branch] gh/nikitaved/15/head -> origin/gh/nikitaved/15/head 2025-12-04T09:20:54.3951481Z * [new branch] gh/nikitaved/15/orig -> origin/gh/nikitaved/15/orig 2025-12-04T09:20:54.3951633Z * [new branch] gh/nikitaved/16/base -> origin/gh/nikitaved/16/base 2025-12-04T09:20:54.3951778Z * [new branch] gh/nikitaved/16/head -> origin/gh/nikitaved/16/head 2025-12-04T09:20:54.3951929Z * [new branch] gh/nikitaved/16/orig -> origin/gh/nikitaved/16/orig 2025-12-04T09:20:54.3952089Z * [new branch] gh/nikitaved/2/base -> origin/gh/nikitaved/2/base 2025-12-04T09:20:54.3952242Z * [new branch] gh/nikitaved/2/head -> origin/gh/nikitaved/2/head 2025-12-04T09:20:54.3956004Z * [new branch] gh/nikitaved/2/orig -> origin/gh/nikitaved/2/orig 2025-12-04T09:20:54.3956159Z * [new branch] gh/nikitaved/4/base -> origin/gh/nikitaved/4/base 2025-12-04T09:20:54.3956319Z * [new branch] gh/nikitaved/4/head -> origin/gh/nikitaved/4/head 2025-12-04T09:20:54.3956673Z * [new branch] gh/nikitaved/4/orig -> origin/gh/nikitaved/4/orig 2025-12-04T09:20:54.3956835Z * [new branch] gh/nikitaved/5/base -> origin/gh/nikitaved/5/base 2025-12-04T09:20:54.3956981Z * [new branch] gh/nikitaved/5/head -> origin/gh/nikitaved/5/head 2025-12-04T09:20:54.3957539Z * [new branch] gh/nikitaved/5/orig -> origin/gh/nikitaved/5/orig 2025-12-04T09:20:54.3959117Z * [new branch] gh/nikitaved/6/base -> origin/gh/nikitaved/6/base 2025-12-04T09:20:54.3959471Z * [new branch] gh/nikitaved/6/head -> origin/gh/nikitaved/6/head 2025-12-04T09:20:54.3959779Z * [new branch] gh/nikitaved/6/orig -> origin/gh/nikitaved/6/orig 2025-12-04T09:20:54.3961030Z * [new branch] gh/nikitaved/8/base -> origin/gh/nikitaved/8/base 2025-12-04T09:20:54.3961393Z * [new branch] gh/nikitaved/8/head -> origin/gh/nikitaved/8/head 2025-12-04T09:20:54.3966181Z * [new branch] gh/nikitaved/8/orig -> origin/gh/nikitaved/8/orig 2025-12-04T09:20:54.3966349Z * [new branch] gh/nikitaved/9/base -> origin/gh/nikitaved/9/base 2025-12-04T09:20:54.3966501Z * [new branch] gh/nikitaved/9/head -> origin/gh/nikitaved/9/head 2025-12-04T09:20:54.3966670Z * [new branch] gh/nikitaved/9/orig -> origin/gh/nikitaved/9/orig 2025-12-04T09:20:54.3966817Z * [new branch] gh/oulgen/10/base -> origin/gh/oulgen/10/base 2025-12-04T09:20:54.3966961Z * [new branch] gh/oulgen/10/head -> origin/gh/oulgen/10/head 2025-12-04T09:20:54.3969172Z * [new branch] gh/oulgen/10/orig -> origin/gh/oulgen/10/orig 2025-12-04T09:20:54.3972922Z * [new branch] gh/oulgen/11/base -> origin/gh/oulgen/11/base 2025-12-04T09:20:54.3973071Z * [new branch] gh/oulgen/11/head -> origin/gh/oulgen/11/head 2025-12-04T09:20:54.3973297Z * [new branch] gh/oulgen/11/orig -> origin/gh/oulgen/11/orig 2025-12-04T09:20:54.3978742Z * [new branch] gh/oulgen/12/base -> origin/gh/oulgen/12/base 2025-12-04T09:20:54.3982347Z * [new branch] gh/oulgen/12/head -> origin/gh/oulgen/12/head 2025-12-04T09:20:54.3986869Z * [new branch] gh/oulgen/12/orig -> origin/gh/oulgen/12/orig 2025-12-04T09:20:54.3991069Z * [new branch] gh/oulgen/13/base -> origin/gh/oulgen/13/base 2025-12-04T09:20:54.3992623Z * [new branch] gh/oulgen/13/head -> origin/gh/oulgen/13/head 2025-12-04T09:20:54.3992795Z * [new branch] gh/oulgen/13/orig -> origin/gh/oulgen/13/orig 2025-12-04T09:20:54.3992931Z * [new branch] gh/oulgen/14/base -> origin/gh/oulgen/14/base 2025-12-04T09:20:54.3993065Z * [new branch] gh/oulgen/14/head -> origin/gh/oulgen/14/head 2025-12-04T09:20:54.3993229Z * [new branch] gh/oulgen/14/orig -> origin/gh/oulgen/14/orig 2025-12-04T09:20:54.3993363Z * [new branch] gh/oulgen/15/base -> origin/gh/oulgen/15/base 2025-12-04T09:20:54.3993497Z * [new branch] gh/oulgen/15/head -> origin/gh/oulgen/15/head 2025-12-04T09:20:54.3993641Z * [new branch] gh/oulgen/15/orig -> origin/gh/oulgen/15/orig 2025-12-04T09:20:54.3993774Z * [new branch] gh/oulgen/16/base -> origin/gh/oulgen/16/base 2025-12-04T09:20:54.3993912Z * [new branch] gh/oulgen/16/head -> origin/gh/oulgen/16/head 2025-12-04T09:20:54.3994045Z * [new branch] gh/oulgen/16/orig -> origin/gh/oulgen/16/orig 2025-12-04T09:20:54.3994182Z * [new branch] gh/oulgen/17/base -> origin/gh/oulgen/17/base 2025-12-04T09:20:54.3994322Z * [new branch] gh/oulgen/17/head -> origin/gh/oulgen/17/head 2025-12-04T09:20:54.3994615Z * [new branch] gh/oulgen/17/orig -> origin/gh/oulgen/17/orig 2025-12-04T09:20:54.3994759Z * [new branch] gh/oulgen/18/base -> origin/gh/oulgen/18/base 2025-12-04T09:20:54.3994894Z * [new branch] gh/oulgen/18/head -> origin/gh/oulgen/18/head 2025-12-04T09:20:54.3995093Z * [new branch] gh/oulgen/18/orig -> origin/gh/oulgen/18/orig 2025-12-04T09:20:54.3995253Z * [new branch] gh/oulgen/19/base -> origin/gh/oulgen/19/base 2025-12-04T09:20:54.3995388Z * [new branch] gh/oulgen/19/head -> origin/gh/oulgen/19/head 2025-12-04T09:20:54.3995531Z * [new branch] gh/oulgen/19/orig -> origin/gh/oulgen/19/orig 2025-12-04T09:20:54.3995667Z * [new branch] gh/oulgen/20/base -> origin/gh/oulgen/20/base 2025-12-04T09:20:54.3995800Z * [new branch] gh/oulgen/20/head -> origin/gh/oulgen/20/head 2025-12-04T09:20:54.3995945Z * [new branch] gh/oulgen/20/orig -> origin/gh/oulgen/20/orig 2025-12-04T09:20:54.3996077Z * [new branch] gh/oulgen/21/base -> origin/gh/oulgen/21/base 2025-12-04T09:20:54.3996210Z * [new branch] gh/oulgen/21/head -> origin/gh/oulgen/21/head 2025-12-04T09:20:54.3996354Z * [new branch] gh/oulgen/21/orig -> origin/gh/oulgen/21/orig 2025-12-04T09:20:54.3996487Z * [new branch] gh/oulgen/22/base -> origin/gh/oulgen/22/base 2025-12-04T09:20:54.3996626Z * [new branch] gh/oulgen/22/head -> origin/gh/oulgen/22/head 2025-12-04T09:20:54.3996757Z * [new branch] gh/oulgen/22/orig -> origin/gh/oulgen/22/orig 2025-12-04T09:20:54.3997949Z * [new branch] gh/oulgen/23/base -> origin/gh/oulgen/23/base 2025-12-04T09:20:54.3998613Z * [new branch] gh/oulgen/23/head -> origin/gh/oulgen/23/head 2025-12-04T09:20:54.3998869Z * [new branch] gh/oulgen/23/orig -> origin/gh/oulgen/23/orig 2025-12-04T09:20:54.3999011Z * [new branch] gh/oulgen/24/base -> origin/gh/oulgen/24/base 2025-12-04T09:20:54.3999161Z * [new branch] gh/oulgen/24/head -> origin/gh/oulgen/24/head 2025-12-04T09:20:54.3999307Z * [new branch] gh/oulgen/24/orig -> origin/gh/oulgen/24/orig 2025-12-04T09:20:54.3999446Z * [new branch] gh/oulgen/25/base -> origin/gh/oulgen/25/base 2025-12-04T09:20:54.3999596Z * [new branch] gh/oulgen/25/head -> origin/gh/oulgen/25/head 2025-12-04T09:20:54.4000195Z * [new branch] gh/oulgen/25/orig -> origin/gh/oulgen/25/orig 2025-12-04T09:20:54.4002902Z * [new branch] gh/oulgen/26/base -> origin/gh/oulgen/26/base 2025-12-04T09:20:54.4008917Z * [new branch] gh/oulgen/26/head -> origin/gh/oulgen/26/head 2025-12-04T09:20:54.4009277Z * [new branch] gh/oulgen/26/orig -> origin/gh/oulgen/26/orig 2025-12-04T09:20:54.4009490Z * [new branch] gh/oulgen/4/base -> origin/gh/oulgen/4/base 2025-12-04T09:20:54.4009682Z * [new branch] gh/oulgen/4/head -> origin/gh/oulgen/4/head 2025-12-04T09:20:54.4009860Z * [new branch] gh/oulgen/4/orig -> origin/gh/oulgen/4/orig 2025-12-04T09:20:54.4010095Z * [new branch] gh/oulgen/7/base -> origin/gh/oulgen/7/base 2025-12-04T09:20:54.4011046Z * [new branch] gh/oulgen/7/head -> origin/gh/oulgen/7/head 2025-12-04T09:20:54.4011717Z * [new branch] gh/oulgen/7/orig -> origin/gh/oulgen/7/orig 2025-12-04T09:20:54.4011894Z * [new branch] gh/oulgen/8/base -> origin/gh/oulgen/8/base 2025-12-04T09:20:54.4012041Z * [new branch] gh/oulgen/8/head -> origin/gh/oulgen/8/head 2025-12-04T09:20:54.4012357Z * [new branch] gh/oulgen/8/orig -> origin/gh/oulgen/8/orig 2025-12-04T09:20:54.4015877Z * [new branch] gh/oulgen/9/base -> origin/gh/oulgen/9/base 2025-12-04T09:20:54.4018228Z * [new branch] gh/oulgen/9/head -> origin/gh/oulgen/9/head 2025-12-04T09:20:54.4018535Z * [new branch] gh/oulgen/9/orig -> origin/gh/oulgen/9/orig 2025-12-04T09:20:54.4018764Z * [new branch] gh/patvig/mtia-serialization -> origin/gh/patvig/mtia-serialization 2025-12-04T09:20:54.4018923Z * [new branch] gh/pearu/108/base -> origin/gh/pearu/108/base 2025-12-04T09:20:54.4019066Z * [new branch] gh/pearu/108/head -> origin/gh/pearu/108/head 2025-12-04T09:20:54.4024785Z * [new branch] gh/pearu/108/orig -> origin/gh/pearu/108/orig 2025-12-04T09:20:54.4024998Z * [new branch] gh/pearu/109/base -> origin/gh/pearu/109/base 2025-12-04T09:20:54.4025389Z * [new branch] gh/pearu/109/head -> origin/gh/pearu/109/head 2025-12-04T09:20:54.4025735Z * [new branch] gh/pearu/109/orig -> origin/gh/pearu/109/orig 2025-12-04T09:20:54.4025888Z * [new branch] gh/pearu/110/base -> origin/gh/pearu/110/base 2025-12-04T09:20:54.4026054Z * [new branch] gh/pearu/110/head -> origin/gh/pearu/110/head 2025-12-04T09:20:54.4026192Z * [new branch] gh/pearu/110/orig -> origin/gh/pearu/110/orig 2025-12-04T09:20:54.4026329Z * [new branch] gh/pearu/111/base -> origin/gh/pearu/111/base 2025-12-04T09:20:54.4026474Z * [new branch] gh/pearu/111/head -> origin/gh/pearu/111/head 2025-12-04T09:20:54.4026609Z * [new branch] gh/pearu/111/orig -> origin/gh/pearu/111/orig 2025-12-04T09:20:54.4031280Z * [new branch] gh/pearu/112/base -> origin/gh/pearu/112/base 2025-12-04T09:20:54.4031484Z * [new branch] gh/pearu/112/head -> origin/gh/pearu/112/head 2025-12-04T09:20:54.4031629Z * [new branch] gh/pearu/112/orig -> origin/gh/pearu/112/orig 2025-12-04T09:20:54.4031777Z * [new branch] gh/pearu/115/base -> origin/gh/pearu/115/base 2025-12-04T09:20:54.4031922Z * [new branch] gh/pearu/115/head -> origin/gh/pearu/115/head 2025-12-04T09:20:54.4032072Z * [new branch] gh/pearu/115/orig -> origin/gh/pearu/115/orig 2025-12-04T09:20:54.4032213Z * [new branch] gh/pearu/116/base -> origin/gh/pearu/116/base 2025-12-04T09:20:54.4032351Z * [new branch] gh/pearu/116/head -> origin/gh/pearu/116/head 2025-12-04T09:20:54.4037566Z * [new branch] gh/pearu/116/orig -> origin/gh/pearu/116/orig 2025-12-04T09:20:54.4037902Z * [new branch] gh/pearu/117/base -> origin/gh/pearu/117/base 2025-12-04T09:20:54.4038095Z * [new branch] gh/pearu/117/head -> origin/gh/pearu/117/head 2025-12-04T09:20:54.4038252Z * [new branch] gh/pearu/117/orig -> origin/gh/pearu/117/orig 2025-12-04T09:20:54.4038459Z * [new branch] gh/pearu/118/base -> origin/gh/pearu/118/base 2025-12-04T09:20:54.4038626Z * [new branch] gh/pearu/118/head -> origin/gh/pearu/118/head 2025-12-04T09:20:54.4038847Z * [new branch] gh/pearu/118/orig -> origin/gh/pearu/118/orig 2025-12-04T09:20:54.4039009Z * [new branch] gh/pearu/119/base -> origin/gh/pearu/119/base 2025-12-04T09:20:54.4039724Z * [new branch] gh/pearu/119/head -> origin/gh/pearu/119/head 2025-12-04T09:20:54.4039907Z * [new branch] gh/pearu/119/orig -> origin/gh/pearu/119/orig 2025-12-04T09:20:54.4040064Z * [new branch] gh/pearu/139/base -> origin/gh/pearu/139/base 2025-12-04T09:20:54.4040456Z * [new branch] gh/pearu/139/head -> origin/gh/pearu/139/head 2025-12-04T09:20:54.4043014Z * [new branch] gh/pearu/139/orig -> origin/gh/pearu/139/orig 2025-12-04T09:20:54.4043330Z * [new branch] gh/pearu/140/base -> origin/gh/pearu/140/base 2025-12-04T09:20:54.4043699Z * [new branch] gh/pearu/140/head -> origin/gh/pearu/140/head 2025-12-04T09:20:54.4043987Z * [new branch] gh/pearu/140/orig -> origin/gh/pearu/140/orig 2025-12-04T09:20:54.4044146Z * [new branch] gh/pearu/142/base -> origin/gh/pearu/142/base 2025-12-04T09:20:54.4044283Z * [new branch] gh/pearu/142/head -> origin/gh/pearu/142/head 2025-12-04T09:20:54.4044429Z * [new branch] gh/pearu/142/orig -> origin/gh/pearu/142/orig 2025-12-04T09:20:54.4046286Z * [new branch] gh/pearu/143/base -> origin/gh/pearu/143/base 2025-12-04T09:20:54.4046495Z * [new branch] gh/pearu/143/head -> origin/gh/pearu/143/head 2025-12-04T09:20:54.4046676Z * [new branch] gh/pearu/143/orig -> origin/gh/pearu/143/orig 2025-12-04T09:20:54.4051379Z * [new branch] gh/pearu/147/base -> origin/gh/pearu/147/base 2025-12-04T09:20:54.4051763Z * [new branch] gh/pearu/147/head -> origin/gh/pearu/147/head 2025-12-04T09:20:54.4052005Z * [new branch] gh/pearu/147/orig -> origin/gh/pearu/147/orig 2025-12-04T09:20:54.4052167Z * [new branch] gh/pearu/149/base -> origin/gh/pearu/149/base 2025-12-04T09:20:54.4052304Z * [new branch] gh/pearu/149/head -> origin/gh/pearu/149/head 2025-12-04T09:20:54.4052511Z * [new branch] gh/pearu/149/orig -> origin/gh/pearu/149/orig 2025-12-04T09:20:54.4057662Z * [new branch] gh/pearu/150/base -> origin/gh/pearu/150/base 2025-12-04T09:20:54.4059785Z * [new branch] gh/pearu/150/head -> origin/gh/pearu/150/head 2025-12-04T09:20:54.4060112Z * [new branch] gh/pearu/150/orig -> origin/gh/pearu/150/orig 2025-12-04T09:20:54.4060285Z * [new branch] gh/pearu/151/base -> origin/gh/pearu/151/base 2025-12-04T09:20:54.4060453Z * [new branch] gh/pearu/151/head -> origin/gh/pearu/151/head 2025-12-04T09:20:54.4060681Z * [new branch] gh/pearu/151/orig -> origin/gh/pearu/151/orig 2025-12-04T09:20:54.4060857Z * [new branch] gh/pearu/152/base -> origin/gh/pearu/152/base 2025-12-04T09:20:54.4060992Z * [new branch] gh/pearu/152/head -> origin/gh/pearu/152/head 2025-12-04T09:20:54.4061132Z * [new branch] gh/pearu/152/orig -> origin/gh/pearu/152/orig 2025-12-04T09:20:54.4061265Z * [new branch] gh/pearu/153/base -> origin/gh/pearu/153/base 2025-12-04T09:20:54.4061535Z * [new branch] gh/pearu/153/head -> origin/gh/pearu/153/head 2025-12-04T09:20:54.4061700Z * [new branch] gh/pearu/153/orig -> origin/gh/pearu/153/orig 2025-12-04T09:20:54.4066159Z * [new branch] gh/pearu/154/base -> origin/gh/pearu/154/base 2025-12-04T09:20:54.4066774Z * [new branch] gh/pearu/154/head -> origin/gh/pearu/154/head 2025-12-04T09:20:54.4067046Z * [new branch] gh/pearu/154/orig -> origin/gh/pearu/154/orig 2025-12-04T09:20:54.4067218Z * [new branch] gh/pearu/155/base -> origin/gh/pearu/155/base 2025-12-04T09:20:54.4067382Z * [new branch] gh/pearu/155/head -> origin/gh/pearu/155/head 2025-12-04T09:20:54.4067558Z * [new branch] gh/pearu/155/orig -> origin/gh/pearu/155/orig 2025-12-04T09:20:54.4068280Z * [new branch] gh/pearu/156/base -> origin/gh/pearu/156/base 2025-12-04T09:20:54.4068808Z * [new branch] gh/pearu/156/head -> origin/gh/pearu/156/head 2025-12-04T09:20:54.4068982Z * [new branch] gh/pearu/156/orig -> origin/gh/pearu/156/orig 2025-12-04T09:20:54.4073605Z * [new branch] gh/pearu/56/base -> origin/gh/pearu/56/base 2025-12-04T09:20:54.4073970Z * [new branch] gh/pearu/56/head -> origin/gh/pearu/56/head 2025-12-04T09:20:54.4074120Z * [new branch] gh/pearu/56/orig -> origin/gh/pearu/56/orig 2025-12-04T09:20:54.4074268Z * [new branch] gh/pearu/97/base -> origin/gh/pearu/97/base 2025-12-04T09:20:54.4074786Z * [new branch] gh/pearu/97/head -> origin/gh/pearu/97/head 2025-12-04T09:20:54.4075328Z * [new branch] gh/pearu/97/orig -> origin/gh/pearu/97/orig 2025-12-04T09:20:54.4077392Z * [new branch] gh/pianpwk/21/base -> origin/gh/pianpwk/21/base 2025-12-04T09:20:54.4077768Z * [new branch] gh/pianpwk/21/head -> origin/gh/pianpwk/21/head 2025-12-04T09:20:54.4079164Z * [new branch] gh/pianpwk/28/base -> origin/gh/pianpwk/28/base 2025-12-04T09:20:54.4079639Z * [new branch] gh/pianpwk/28/head -> origin/gh/pianpwk/28/head 2025-12-04T09:20:54.4080833Z * [new branch] gh/pianpwk/28/orig -> origin/gh/pianpwk/28/orig 2025-12-04T09:20:54.4081471Z * [new branch] gh/pianpwk/29/base -> origin/gh/pianpwk/29/base 2025-12-04T09:20:54.4086641Z * [new branch] gh/pianpwk/29/head -> origin/gh/pianpwk/29/head 2025-12-04T09:20:54.4086825Z * [new branch] gh/pianpwk/29/orig -> origin/gh/pianpwk/29/orig 2025-12-04T09:20:54.4087377Z * [new branch] gh/pianpwk/30/base -> origin/gh/pianpwk/30/base 2025-12-04T09:20:54.4087567Z * [new branch] gh/pianpwk/30/head -> origin/gh/pianpwk/30/head 2025-12-04T09:20:54.4087735Z * [new branch] gh/pianpwk/30/orig -> origin/gh/pianpwk/30/orig 2025-12-04T09:20:54.4087888Z * [new branch] gh/pianpwk/31/base -> origin/gh/pianpwk/31/base 2025-12-04T09:20:54.4088040Z * [new branch] gh/pianpwk/31/head -> origin/gh/pianpwk/31/head 2025-12-04T09:20:54.4090832Z * [new branch] gh/pianpwk/31/orig -> origin/gh/pianpwk/31/orig 2025-12-04T09:20:54.4091039Z * [new branch] gh/pianpwk/32/base -> origin/gh/pianpwk/32/base 2025-12-04T09:20:54.4091198Z * [new branch] gh/pianpwk/32/head -> origin/gh/pianpwk/32/head 2025-12-04T09:20:54.4091430Z * [new branch] gh/pianpwk/32/orig -> origin/gh/pianpwk/32/orig 2025-12-04T09:20:54.4091593Z * [new branch] gh/pianpwk/33/base -> origin/gh/pianpwk/33/base 2025-12-04T09:20:54.4097153Z * [new branch] gh/pianpwk/33/head -> origin/gh/pianpwk/33/head 2025-12-04T09:20:54.4099323Z * [new branch] gh/pianpwk/33/orig -> origin/gh/pianpwk/33/orig 2025-12-04T09:20:54.4099478Z * [new branch] gh/pianpwk/34/base -> origin/gh/pianpwk/34/base 2025-12-04T09:20:54.4099645Z * [new branch] gh/pianpwk/34/head -> origin/gh/pianpwk/34/head 2025-12-04T09:20:54.4099816Z * [new branch] gh/pianpwk/34/orig -> origin/gh/pianpwk/34/orig 2025-12-04T09:20:54.4099970Z * [new branch] gh/pianpwk/35/base -> origin/gh/pianpwk/35/base 2025-12-04T09:20:54.4100119Z * [new branch] gh/pianpwk/35/head -> origin/gh/pianpwk/35/head 2025-12-04T09:20:54.4100285Z * [new branch] gh/pianpwk/35/orig -> origin/gh/pianpwk/35/orig 2025-12-04T09:20:54.4100758Z * [new branch] gh/rec/141/base -> origin/gh/rec/141/base 2025-12-04T09:20:54.4100923Z * [new branch] gh/rec/141/head -> origin/gh/rec/141/head 2025-12-04T09:20:54.4104472Z * [new branch] gh/rec/153/base -> origin/gh/rec/153/base 2025-12-04T09:20:54.4104636Z * [new branch] gh/rec/153/head -> origin/gh/rec/153/head 2025-12-04T09:20:54.4104785Z * [new branch] gh/rec/153/orig -> origin/gh/rec/153/orig 2025-12-04T09:20:54.4104989Z * [new branch] gh/rec/154/base -> origin/gh/rec/154/base 2025-12-04T09:20:54.4105133Z * [new branch] gh/rec/154/head -> origin/gh/rec/154/head 2025-12-04T09:20:54.4108191Z * [new branch] gh/rec/154/orig -> origin/gh/rec/154/orig 2025-12-04T09:20:54.4108545Z * [new branch] gh/rec/164/base -> origin/gh/rec/164/base 2025-12-04T09:20:54.4108713Z * [new branch] gh/rec/164/head -> origin/gh/rec/164/head 2025-12-04T09:20:54.4108853Z * [new branch] gh/rec/164/orig -> origin/gh/rec/164/orig 2025-12-04T09:20:54.4114158Z * [new branch] gh/rec/166/base -> origin/gh/rec/166/base 2025-12-04T09:20:54.4114417Z * [new branch] gh/rec/166/head -> origin/gh/rec/166/head 2025-12-04T09:20:54.4114552Z * [new branch] gh/rec/166/orig -> origin/gh/rec/166/orig 2025-12-04T09:20:54.4114689Z * [new branch] gh/rec/167/base -> origin/gh/rec/167/base 2025-12-04T09:20:54.4114827Z * [new branch] gh/rec/167/head -> origin/gh/rec/167/head 2025-12-04T09:20:54.4114955Z * [new branch] gh/rec/167/orig -> origin/gh/rec/167/orig 2025-12-04T09:20:54.4115093Z * [new branch] gh/rec/168/base -> origin/gh/rec/168/base 2025-12-04T09:20:54.4118489Z * [new branch] gh/rec/168/head -> origin/gh/rec/168/head 2025-12-04T09:20:54.4118777Z * [new branch] gh/rec/168/orig -> origin/gh/rec/168/orig 2025-12-04T09:20:54.4118929Z * [new branch] gh/rec/169/base -> origin/gh/rec/169/base 2025-12-04T09:20:54.4119071Z * [new branch] gh/rec/169/head -> origin/gh/rec/169/head 2025-12-04T09:20:54.4119203Z * [new branch] gh/rec/169/orig -> origin/gh/rec/169/orig 2025-12-04T09:20:54.4119352Z * [new branch] gh/rec/170/base -> origin/gh/rec/170/base 2025-12-04T09:20:54.4120037Z * [new branch] gh/rec/170/head -> origin/gh/rec/170/head 2025-12-04T09:20:54.4120797Z * [new branch] gh/rec/170/orig -> origin/gh/rec/170/orig 2025-12-04T09:20:54.4122007Z * [new branch] gh/rec/171/base -> origin/gh/rec/171/base 2025-12-04T09:20:54.4122535Z * [new branch] gh/rec/171/head -> origin/gh/rec/171/head 2025-12-04T09:20:54.4127480Z * [new branch] gh/rec/171/orig -> origin/gh/rec/171/orig 2025-12-04T09:20:54.4127798Z * [new branch] gh/rec/172/base -> origin/gh/rec/172/base 2025-12-04T09:20:54.4127978Z * [new branch] gh/rec/172/head -> origin/gh/rec/172/head 2025-12-04T09:20:54.4128155Z * [new branch] gh/rec/172/orig -> origin/gh/rec/172/orig 2025-12-04T09:20:54.4128326Z * [new branch] gh/rec/173/base -> origin/gh/rec/173/base 2025-12-04T09:20:54.4128498Z * [new branch] gh/rec/173/head -> origin/gh/rec/173/head 2025-12-04T09:20:54.4136230Z * [new branch] gh/rec/173/orig -> origin/gh/rec/173/orig 2025-12-04T09:20:54.4136825Z * [new branch] gh/rec/174/base -> origin/gh/rec/174/base 2025-12-04T09:20:54.4137015Z * [new branch] gh/rec/174/head -> origin/gh/rec/174/head 2025-12-04T09:20:54.4137148Z * [new branch] gh/rec/174/orig -> origin/gh/rec/174/orig 2025-12-04T09:20:54.4137501Z * [new branch] gh/rec/175/base -> origin/gh/rec/175/base 2025-12-04T09:20:54.4137641Z * [new branch] gh/rec/175/head -> origin/gh/rec/175/head 2025-12-04T09:20:54.4137773Z * [new branch] gh/rec/175/orig -> origin/gh/rec/175/orig 2025-12-04T09:20:54.4137970Z * [new branch] gh/rec/176/base -> origin/gh/rec/176/base 2025-12-04T09:20:54.4138105Z * [new branch] gh/rec/176/head -> origin/gh/rec/176/head 2025-12-04T09:20:54.4138236Z * [new branch] gh/rec/176/orig -> origin/gh/rec/176/orig 2025-12-04T09:20:54.4141001Z * [new branch] gh/rec/177/base -> origin/gh/rec/177/base 2025-12-04T09:20:54.4141195Z * [new branch] gh/rec/177/head -> origin/gh/rec/177/head 2025-12-04T09:20:54.4141360Z * [new branch] gh/rec/177/orig -> origin/gh/rec/177/orig 2025-12-04T09:20:54.4141595Z * [new branch] gh/robert-hardwick/3/base -> origin/gh/robert-hardwick/3/base 2025-12-04T09:20:54.4141843Z * [new branch] gh/robert-hardwick/3/head -> origin/gh/robert-hardwick/3/head 2025-12-04T09:20:54.4142030Z * [new branch] gh/robert-hardwick/3/orig -> origin/gh/robert-hardwick/3/orig 2025-12-04T09:20:54.4147493Z * [new branch] gh/robert-hardwick/4/base -> origin/gh/robert-hardwick/4/base 2025-12-04T09:20:54.4147700Z * [new branch] gh/robert-hardwick/4/head -> origin/gh/robert-hardwick/4/head 2025-12-04T09:20:54.4148178Z * [new branch] gh/robert-hardwick/4/orig -> origin/gh/robert-hardwick/4/orig 2025-12-04T09:20:54.4148390Z * [new branch] gh/robert-hardwick/5/base -> origin/gh/robert-hardwick/5/base 2025-12-04T09:20:54.4148574Z * [new branch] gh/robert-hardwick/5/head -> origin/gh/robert-hardwick/5/head 2025-12-04T09:20:54.4148752Z * [new branch] gh/robert-hardwick/5/orig -> origin/gh/robert-hardwick/5/orig 2025-12-04T09:20:54.4151312Z * [new branch] gh/robert-hardwick/6/base -> origin/gh/robert-hardwick/6/base 2025-12-04T09:20:54.4151931Z * [new branch] gh/robert-hardwick/6/head -> origin/gh/robert-hardwick/6/head 2025-12-04T09:20:54.4152143Z * [new branch] gh/robert-hardwick/6/orig -> origin/gh/robert-hardwick/6/orig 2025-12-04T09:20:54.4152333Z * [new branch] gh/robert-hardwick/7/base -> origin/gh/robert-hardwick/7/base 2025-12-04T09:20:54.4152510Z * [new branch] gh/robert-hardwick/7/head -> origin/gh/robert-hardwick/7/head 2025-12-04T09:20:54.4152675Z * [new branch] gh/robert-hardwick/7/orig -> origin/gh/robert-hardwick/7/orig 2025-12-04T09:20:54.4152847Z * [new branch] gh/robert-hardwick/8/base -> origin/gh/robert-hardwick/8/base 2025-12-04T09:20:54.4157823Z * [new branch] gh/robert-hardwick/8/head -> origin/gh/robert-hardwick/8/head 2025-12-04T09:20:54.4158051Z * [new branch] gh/robert-hardwick/8/orig -> origin/gh/robert-hardwick/8/orig 2025-12-04T09:20:54.4158231Z * [new branch] gh/robert-hardwick/9/base -> origin/gh/robert-hardwick/9/base 2025-12-04T09:20:54.4158401Z * [new branch] gh/robert-hardwick/9/head -> origin/gh/robert-hardwick/9/head 2025-12-04T09:20:54.4158587Z * [new branch] gh/robert-hardwick/9/orig -> origin/gh/robert-hardwick/9/orig 2025-12-04T09:20:54.4159199Z * [new branch] gh/rtimpe/1/base -> origin/gh/rtimpe/1/base 2025-12-04T09:20:54.4159348Z * [new branch] gh/rtimpe/1/head -> origin/gh/rtimpe/1/head 2025-12-04T09:20:54.4159495Z * [new branch] gh/rtimpe/2/base -> origin/gh/rtimpe/2/base 2025-12-04T09:20:54.4159632Z * [new branch] gh/rtimpe/2/head -> origin/gh/rtimpe/2/head 2025-12-04T09:20:54.4159782Z * [new branch] gh/rtimpe/22/base -> origin/gh/rtimpe/22/base 2025-12-04T09:20:54.4160374Z * [new branch] gh/rtimpe/22/head -> origin/gh/rtimpe/22/head 2025-12-04T09:20:54.4161216Z * [new branch] gh/rtimpe/22/orig -> origin/gh/rtimpe/22/orig 2025-12-04T09:20:54.4162359Z * [new branch] gh/rtimpe/23/base -> origin/gh/rtimpe/23/base 2025-12-04T09:20:54.4168313Z * [new branch] gh/rtimpe/23/head -> origin/gh/rtimpe/23/head 2025-12-04T09:20:54.4168664Z * [new branch] gh/rtimpe/23/orig -> origin/gh/rtimpe/23/orig 2025-12-04T09:20:54.4168833Z * [new branch] gh/rtimpe/24/base -> origin/gh/rtimpe/24/base 2025-12-04T09:20:54.4168969Z * [new branch] gh/rtimpe/24/head -> origin/gh/rtimpe/24/head 2025-12-04T09:20:54.4169112Z * [new branch] gh/rtimpe/24/orig -> origin/gh/rtimpe/24/orig 2025-12-04T09:20:54.4169379Z * [new branch] gh/rtimpe/25/base -> origin/gh/rtimpe/25/base 2025-12-04T09:20:54.4174602Z * [new branch] gh/rtimpe/25/head -> origin/gh/rtimpe/25/head 2025-12-04T09:20:54.4174923Z * [new branch] gh/rtimpe/25/orig -> origin/gh/rtimpe/25/orig 2025-12-04T09:20:54.4175139Z * [new branch] gh/rtimpe/26/base -> origin/gh/rtimpe/26/base 2025-12-04T09:20:54.4175309Z * [new branch] gh/rtimpe/26/head -> origin/gh/rtimpe/26/head 2025-12-04T09:20:54.4175450Z * [new branch] gh/rtimpe/26/orig -> origin/gh/rtimpe/26/orig 2025-12-04T09:20:54.4175599Z * [new branch] gh/rtimpe/27/base -> origin/gh/rtimpe/27/base 2025-12-04T09:20:54.4175744Z * [new branch] gh/rtimpe/27/head -> origin/gh/rtimpe/27/head 2025-12-04T09:20:54.4175886Z * [new branch] gh/rtimpe/27/orig -> origin/gh/rtimpe/27/orig 2025-12-04T09:20:54.4176038Z * [new branch] gh/rtimpe/28/base -> origin/gh/rtimpe/28/base 2025-12-04T09:20:54.4176185Z * [new branch] gh/rtimpe/28/head -> origin/gh/rtimpe/28/head 2025-12-04T09:20:54.4176326Z * [new branch] gh/rtimpe/28/orig -> origin/gh/rtimpe/28/orig 2025-12-04T09:20:54.4176480Z * [new branch] gh/rtimpe/29/base -> origin/gh/rtimpe/29/base 2025-12-04T09:20:54.4180998Z * [new branch] gh/rtimpe/29/head -> origin/gh/rtimpe/29/head 2025-12-04T09:20:54.4181198Z * [new branch] gh/rtimpe/29/orig -> origin/gh/rtimpe/29/orig 2025-12-04T09:20:54.4181355Z * [new branch] gh/rtimpe/3/base -> origin/gh/rtimpe/3/base 2025-12-04T09:20:54.4181504Z * [new branch] gh/rtimpe/3/head -> origin/gh/rtimpe/3/head 2025-12-04T09:20:54.4181646Z * [new branch] gh/rtimpe/30/base -> origin/gh/rtimpe/30/base 2025-12-04T09:20:54.4181782Z * [new branch] gh/rtimpe/30/head -> origin/gh/rtimpe/30/head 2025-12-04T09:20:54.4187579Z * [new branch] gh/rtimpe/30/orig -> origin/gh/rtimpe/30/orig 2025-12-04T09:20:54.4187914Z * [new branch] gh/rtimpe/31/base -> origin/gh/rtimpe/31/base 2025-12-04T09:20:54.4188086Z * [new branch] gh/rtimpe/31/head -> origin/gh/rtimpe/31/head 2025-12-04T09:20:54.4188279Z * [new branch] gh/rtimpe/31/orig -> origin/gh/rtimpe/31/orig 2025-12-04T09:20:54.4188449Z * [new branch] gh/rtimpe/32/base -> origin/gh/rtimpe/32/base 2025-12-04T09:20:54.4188605Z * [new branch] gh/rtimpe/32/head -> origin/gh/rtimpe/32/head 2025-12-04T09:20:54.4189282Z * [new branch] gh/rtimpe/32/orig -> origin/gh/rtimpe/32/orig 2025-12-04T09:20:54.4189636Z * [new branch] gh/rtimpe/33/base -> origin/gh/rtimpe/33/base 2025-12-04T09:20:54.4190008Z * [new branch] gh/rtimpe/33/head -> origin/gh/rtimpe/33/head 2025-12-04T09:20:54.4190335Z * [new branch] gh/rtimpe/33/orig -> origin/gh/rtimpe/33/orig 2025-12-04T09:20:54.4190496Z * [new branch] gh/rtimpe/34/base -> origin/gh/rtimpe/34/base 2025-12-04T09:20:54.4190646Z * [new branch] gh/rtimpe/34/head -> origin/gh/rtimpe/34/head 2025-12-04T09:20:54.4193873Z * [new branch] gh/rtimpe/34/orig -> origin/gh/rtimpe/34/orig 2025-12-04T09:20:54.4194064Z * [new branch] gh/rtimpe/35/base -> origin/gh/rtimpe/35/base 2025-12-04T09:20:54.4194574Z * [new branch] gh/rtimpe/35/head -> origin/gh/rtimpe/35/head 2025-12-04T09:20:54.4194742Z * [new branch] gh/rtimpe/35/orig -> origin/gh/rtimpe/35/orig 2025-12-04T09:20:54.4194910Z * [new branch] gh/rtimpe/4/base -> origin/gh/rtimpe/4/base 2025-12-04T09:20:54.4195051Z * [new branch] gh/rtimpe/4/head -> origin/gh/rtimpe/4/head 2025-12-04T09:20:54.4199790Z * [new branch] gh/ruisizhang123/1/base -> origin/gh/ruisizhang123/1/base 2025-12-04T09:20:54.4200150Z * [new branch] gh/ruisizhang123/1/head -> origin/gh/ruisizhang123/1/head 2025-12-04T09:20:54.4200328Z * [new branch] gh/ruisizhang123/1/orig -> origin/gh/ruisizhang123/1/orig 2025-12-04T09:20:54.4200525Z * [new branch] gh/ruisizhang123/4/base -> origin/gh/ruisizhang123/4/base 2025-12-04T09:20:54.4201290Z * [new branch] gh/ruisizhang123/4/head -> origin/gh/ruisizhang123/4/head 2025-12-04T09:20:54.4201502Z * [new branch] gh/ruisizhang123/4/orig -> origin/gh/ruisizhang123/4/orig 2025-12-04T09:20:54.4202075Z * [new branch] gh/ruisizhang123/5/base -> origin/gh/ruisizhang123/5/base 2025-12-04T09:20:54.4202265Z * [new branch] gh/ruisizhang123/5/head -> origin/gh/ruisizhang123/5/head 2025-12-04T09:20:54.4202455Z * [new branch] gh/ruisizhang123/5/orig -> origin/gh/ruisizhang123/5/orig 2025-12-04T09:20:54.4202618Z * [new branch] gh/ruisizhang123/6/base -> origin/gh/ruisizhang123/6/base 2025-12-04T09:20:54.4206632Z * [new branch] gh/ruisizhang123/6/head -> origin/gh/ruisizhang123/6/head 2025-12-04T09:20:54.4206947Z * [new branch] gh/ruisizhang123/6/orig -> origin/gh/ruisizhang123/6/orig 2025-12-04T09:20:54.4207129Z * [new branch] gh/ruisizhang123/7/base -> origin/gh/ruisizhang123/7/base 2025-12-04T09:20:54.4207302Z * [new branch] gh/ruisizhang123/7/head -> origin/gh/ruisizhang123/7/head 2025-12-04T09:20:54.4207466Z * [new branch] gh/ruisizhang123/7/orig -> origin/gh/ruisizhang123/7/orig 2025-12-04T09:20:54.4207768Z * [new branch] gh/ruisizhang123/8/base -> origin/gh/ruisizhang123/8/base 2025-12-04T09:20:54.4208219Z * [new branch] gh/ruisizhang123/8/head -> origin/gh/ruisizhang123/8/head 2025-12-04T09:20:54.4214337Z * [new branch] gh/ruisizhang123/8/orig -> origin/gh/ruisizhang123/8/orig 2025-12-04T09:20:54.4218818Z * [new branch] gh/ruisizhang123/9/base -> origin/gh/ruisizhang123/9/base 2025-12-04T09:20:54.4223757Z * [new branch] gh/ruisizhang123/9/head -> origin/gh/ruisizhang123/9/head 2025-12-04T09:20:54.4227622Z * [new branch] gh/ruisizhang123/9/orig -> origin/gh/ruisizhang123/9/orig 2025-12-04T09:20:54.4233997Z * [new branch] gh/seemethere/52/base -> origin/gh/seemethere/52/base 2025-12-04T09:20:54.4238560Z * [new branch] gh/seemethere/52/head -> origin/gh/seemethere/52/head 2025-12-04T09:20:54.4238760Z * [new branch] gh/seemethere/52/orig -> origin/gh/seemethere/52/orig 2025-12-04T09:20:54.4239049Z * [new branch] gh/seemethere/53/base -> origin/gh/seemethere/53/base 2025-12-04T09:20:54.4239236Z * [new branch] gh/seemethere/53/head -> origin/gh/seemethere/53/head 2025-12-04T09:20:54.4239656Z * [new branch] gh/seemethere/53/orig -> origin/gh/seemethere/53/orig 2025-12-04T09:20:54.4239825Z * [new branch] gh/seemethere/54/base -> origin/gh/seemethere/54/base 2025-12-04T09:20:54.4239975Z * [new branch] gh/seemethere/54/head -> origin/gh/seemethere/54/head 2025-12-04T09:20:54.4240202Z * [new branch] gh/seemethere/54/orig -> origin/gh/seemethere/54/orig 2025-12-04T09:20:54.4240364Z * [new branch] gh/seemethere/55/base -> origin/gh/seemethere/55/base 2025-12-04T09:20:54.4240514Z * [new branch] gh/seemethere/55/head -> origin/gh/seemethere/55/head 2025-12-04T09:20:54.4240666Z * [new branch] gh/seemethere/55/orig -> origin/gh/seemethere/55/orig 2025-12-04T09:20:54.4240822Z * [new branch] gh/seemethere/59/base -> origin/gh/seemethere/59/base 2025-12-04T09:20:54.4240975Z * [new branch] gh/seemethere/59/head -> origin/gh/seemethere/59/head 2025-12-04T09:20:54.4241130Z * [new branch] gh/seemethere/59/orig -> origin/gh/seemethere/59/orig 2025-12-04T09:20:54.4241279Z * [new branch] gh/seemethere/62/base -> origin/gh/seemethere/62/base 2025-12-04T09:20:54.4241432Z * [new branch] gh/seemethere/62/head -> origin/gh/seemethere/62/head 2025-12-04T09:20:54.4241589Z * [new branch] gh/seemethere/62/orig -> origin/gh/seemethere/62/orig 2025-12-04T09:20:54.4241951Z * [new branch] gh/seemethere/63/base -> origin/gh/seemethere/63/base 2025-12-04T09:20:54.4242147Z * [new branch] gh/seemethere/63/head -> origin/gh/seemethere/63/head 2025-12-04T09:20:54.4242301Z * [new branch] gh/seemethere/63/orig -> origin/gh/seemethere/63/orig 2025-12-04T09:20:54.4242452Z * [new branch] gh/seemethere/71/base -> origin/gh/seemethere/71/base 2025-12-04T09:20:54.4242619Z * [new branch] gh/seemethere/71/head -> origin/gh/seemethere/71/head 2025-12-04T09:20:54.4242766Z * [new branch] gh/seemethere/71/orig -> origin/gh/seemethere/71/orig 2025-12-04T09:20:54.4242920Z * [new branch] gh/seemethere/72/base -> origin/gh/seemethere/72/base 2025-12-04T09:20:54.4243069Z * [new branch] gh/seemethere/72/head -> origin/gh/seemethere/72/head 2025-12-04T09:20:54.4243216Z * [new branch] gh/seemethere/72/orig -> origin/gh/seemethere/72/orig 2025-12-04T09:20:54.4243368Z * [new branch] gh/seemethere/73/base -> origin/gh/seemethere/73/base 2025-12-04T09:20:54.4243514Z * [new branch] gh/seemethere/73/head -> origin/gh/seemethere/73/head 2025-12-04T09:20:54.4243680Z * [new branch] gh/seemethere/73/orig -> origin/gh/seemethere/73/orig 2025-12-04T09:20:54.4243826Z * [new branch] gh/seemethere/74/base -> origin/gh/seemethere/74/base 2025-12-04T09:20:54.4243975Z * [new branch] gh/seemethere/74/head -> origin/gh/seemethere/74/head 2025-12-04T09:20:54.4244128Z * [new branch] gh/seemethere/74/orig -> origin/gh/seemethere/74/orig 2025-12-04T09:20:54.4244275Z * [new branch] gh/seemethere/75/base -> origin/gh/seemethere/75/base 2025-12-04T09:20:54.4244423Z * [new branch] gh/seemethere/75/head -> origin/gh/seemethere/75/head 2025-12-04T09:20:54.4244577Z * [new branch] gh/seemethere/75/orig -> origin/gh/seemethere/75/orig 2025-12-04T09:20:54.4244724Z * [new branch] gh/seemethere/76/base -> origin/gh/seemethere/76/base 2025-12-04T09:20:54.4244879Z * [new branch] gh/seemethere/76/head -> origin/gh/seemethere/76/head 2025-12-04T09:20:54.4245025Z * [new branch] gh/seemethere/76/orig -> origin/gh/seemethere/76/orig 2025-12-04T09:20:54.4245245Z * [new branch] gh/shunting314/145/base -> origin/gh/shunting314/145/base 2025-12-04T09:20:54.4249545Z * [new branch] gh/shunting314/145/head -> origin/gh/shunting314/145/head 2025-12-04T09:20:54.4249742Z * [new branch] gh/shunting314/145/orig -> origin/gh/shunting314/145/orig 2025-12-04T09:20:54.4250119Z * [new branch] gh/shunting314/176/base -> origin/gh/shunting314/176/base 2025-12-04T09:20:54.4250281Z * [new branch] gh/shunting314/176/head -> origin/gh/shunting314/176/head 2025-12-04T09:20:54.4250438Z * [new branch] gh/shunting314/176/orig -> origin/gh/shunting314/176/orig 2025-12-04T09:20:54.4250621Z * [new branch] gh/shunting314/249/base -> origin/gh/shunting314/249/base 2025-12-04T09:20:54.4250776Z * [new branch] gh/shunting314/249/head -> origin/gh/shunting314/249/head 2025-12-04T09:20:54.4250938Z * [new branch] gh/shunting314/249/orig -> origin/gh/shunting314/249/orig 2025-12-04T09:20:54.4251096Z * [new branch] gh/shunting314/253/base -> origin/gh/shunting314/253/base 2025-12-04T09:20:54.4251247Z * [new branch] gh/shunting314/253/head -> origin/gh/shunting314/253/head 2025-12-04T09:20:54.4256482Z * [new branch] gh/shunting314/253/orig -> origin/gh/shunting314/253/orig 2025-12-04T09:20:54.4256860Z * [new branch] gh/shunting314/256/base -> origin/gh/shunting314/256/base 2025-12-04T09:20:54.4257051Z * [new branch] gh/shunting314/256/head -> origin/gh/shunting314/256/head 2025-12-04T09:20:54.4257243Z * [new branch] gh/shunting314/256/orig -> origin/gh/shunting314/256/orig 2025-12-04T09:20:54.4257408Z * [new branch] gh/shunting314/257/base -> origin/gh/shunting314/257/base 2025-12-04T09:20:54.4259890Z * [new branch] gh/shunting314/257/head -> origin/gh/shunting314/257/head 2025-12-04T09:20:54.4260077Z * [new branch] gh/shunting314/257/orig -> origin/gh/shunting314/257/orig 2025-12-04T09:20:54.4260239Z * [new branch] gh/shunting314/258/base -> origin/gh/shunting314/258/base 2025-12-04T09:20:54.4260524Z * [new branch] gh/shunting314/258/head -> origin/gh/shunting314/258/head 2025-12-04T09:20:54.4260752Z * [new branch] gh/shunting314/258/orig -> origin/gh/shunting314/258/orig 2025-12-04T09:20:54.4260910Z * [new branch] gh/shunting314/259/base -> origin/gh/shunting314/259/base 2025-12-04T09:20:54.4261236Z * [new branch] gh/shunting314/259/head -> origin/gh/shunting314/259/head 2025-12-04T09:20:54.4264701Z * [new branch] gh/shunting314/259/orig -> origin/gh/shunting314/259/orig 2025-12-04T09:20:54.4264888Z * [new branch] gh/shunting314/260/base -> origin/gh/shunting314/260/base 2025-12-04T09:20:54.4265109Z * [new branch] gh/shunting314/260/head -> origin/gh/shunting314/260/head 2025-12-04T09:20:54.4265287Z * [new branch] gh/shunting314/260/orig -> origin/gh/shunting314/260/orig 2025-12-04T09:20:54.4265456Z * [new branch] gh/shunting314/261/base -> origin/gh/shunting314/261/base 2025-12-04T09:20:54.4265611Z * [new branch] gh/shunting314/261/head -> origin/gh/shunting314/261/head 2025-12-04T09:20:54.4268919Z * [new branch] gh/shunting314/261/orig -> origin/gh/shunting314/261/orig 2025-12-04T09:20:54.4269667Z * [new branch] gh/shunting314/262/base -> origin/gh/shunting314/262/base 2025-12-04T09:20:54.4269856Z * [new branch] gh/shunting314/262/head -> origin/gh/shunting314/262/head 2025-12-04T09:20:54.4270125Z * [new branch] gh/shunting314/262/orig -> origin/gh/shunting314/262/orig 2025-12-04T09:20:54.4270290Z * [new branch] gh/shunting314/263/base -> origin/gh/shunting314/263/base 2025-12-04T09:20:54.4272790Z * [new branch] gh/shunting314/263/head -> origin/gh/shunting314/263/head 2025-12-04T09:20:54.4272959Z * [new branch] gh/shunting314/263/orig -> origin/gh/shunting314/263/orig 2025-12-04T09:20:54.4273133Z * [new branch] gh/shunting314/264/base -> origin/gh/shunting314/264/base 2025-12-04T09:20:54.4273405Z * [new branch] gh/shunting314/264/head -> origin/gh/shunting314/264/head 2025-12-04T09:20:54.4273573Z * [new branch] gh/shunting314/264/orig -> origin/gh/shunting314/264/orig 2025-12-04T09:20:54.4273745Z * [new branch] gh/shunting314/265/base -> origin/gh/shunting314/265/base 2025-12-04T09:20:54.4278251Z * [new branch] gh/shunting314/265/head -> origin/gh/shunting314/265/head 2025-12-04T09:20:54.4278453Z * [new branch] gh/shunting314/265/orig -> origin/gh/shunting314/265/orig 2025-12-04T09:20:54.4278615Z * [new branch] gh/shunting314/266/base -> origin/gh/shunting314/266/base 2025-12-04T09:20:54.4278802Z * [new branch] gh/shunting314/266/head -> origin/gh/shunting314/266/head 2025-12-04T09:20:54.4278959Z * [new branch] gh/shunting314/266/orig -> origin/gh/shunting314/266/orig 2025-12-04T09:20:54.4279162Z * [new branch] gh/shunting314/267/base -> origin/gh/shunting314/267/base 2025-12-04T09:20:54.4279488Z * [new branch] gh/shunting314/267/head -> origin/gh/shunting314/267/head 2025-12-04T09:20:54.4282726Z * [new branch] gh/shunting314/267/orig -> origin/gh/shunting314/267/orig 2025-12-04T09:20:54.4282922Z * [new branch] gh/shunting314/268/base -> origin/gh/shunting314/268/base 2025-12-04T09:20:54.4283093Z * [new branch] gh/shunting314/268/head -> origin/gh/shunting314/268/head 2025-12-04T09:20:54.4283479Z * [new branch] gh/shunting314/268/orig -> origin/gh/shunting314/268/orig 2025-12-04T09:20:54.4287342Z * [new branch] gh/shunting314/269/base -> origin/gh/shunting314/269/base 2025-12-04T09:20:54.4287533Z * [new branch] gh/shunting314/269/head -> origin/gh/shunting314/269/head 2025-12-04T09:20:54.4288000Z * [new branch] gh/shunting314/269/orig -> origin/gh/shunting314/269/orig 2025-12-04T09:20:54.4288228Z * [new branch] gh/silverguo/1/base -> origin/gh/silverguo/1/base 2025-12-04T09:20:54.4288605Z * [new branch] gh/silverguo/1/head -> origin/gh/silverguo/1/head 2025-12-04T09:20:54.4288915Z * [new branch] gh/silverguo/2/base -> origin/gh/silverguo/2/base 2025-12-04T09:20:54.4289449Z * [new branch] gh/silverguo/2/head -> origin/gh/silverguo/2/head 2025-12-04T09:20:54.4291151Z * [new branch] gh/silverguo/3/base -> origin/gh/silverguo/3/base 2025-12-04T09:20:54.4291352Z * [new branch] gh/silverguo/3/head -> origin/gh/silverguo/3/head 2025-12-04T09:20:54.4295668Z * [new branch] gh/silverguo/4/base -> origin/gh/silverguo/4/base 2025-12-04T09:20:54.4300091Z * [new branch] gh/silverguo/4/head -> origin/gh/silverguo/4/head 2025-12-04T09:20:54.4304574Z * [new branch] gh/slayton58/39/base -> origin/gh/slayton58/39/base 2025-12-04T09:20:54.4308931Z * [new branch] gh/slayton58/39/head -> origin/gh/slayton58/39/head 2025-12-04T09:20:54.4309373Z * [new branch] gh/slayton58/39/orig -> origin/gh/slayton58/39/orig 2025-12-04T09:20:54.4309551Z * [new branch] gh/slayton58/42/base -> origin/gh/slayton58/42/base 2025-12-04T09:20:54.4309723Z * [new branch] gh/slayton58/42/head -> origin/gh/slayton58/42/head 2025-12-04T09:20:54.4309889Z * [new branch] gh/slayton58/42/orig -> origin/gh/slayton58/42/orig 2025-12-04T09:20:54.4310093Z * [new branch] gh/slayton58/43/base -> origin/gh/slayton58/43/base 2025-12-04T09:20:54.4310904Z * [new branch] gh/slayton58/43/head -> origin/gh/slayton58/43/head 2025-12-04T09:20:54.4314956Z * [new branch] gh/slayton58/43/orig -> origin/gh/slayton58/43/orig 2025-12-04T09:20:54.4315115Z * [new branch] gh/slayton58/44/base -> origin/gh/slayton58/44/base 2025-12-04T09:20:54.4315489Z * [new branch] gh/slayton58/44/head -> origin/gh/slayton58/44/head 2025-12-04T09:20:54.4315660Z * [new branch] gh/slayton58/44/orig -> origin/gh/slayton58/44/orig 2025-12-04T09:20:54.4315820Z * [new branch] gh/slayton58/45/base -> origin/gh/slayton58/45/base 2025-12-04T09:20:54.4315995Z * [new branch] gh/slayton58/45/head -> origin/gh/slayton58/45/head 2025-12-04T09:20:54.4316149Z * [new branch] gh/slayton58/45/orig -> origin/gh/slayton58/45/orig 2025-12-04T09:20:54.4316312Z * [new branch] gh/slayton58/46/base -> origin/gh/slayton58/46/base 2025-12-04T09:20:54.4316462Z * [new branch] gh/slayton58/46/head -> origin/gh/slayton58/46/head 2025-12-04T09:20:54.4316622Z * [new branch] gh/slayton58/46/orig -> origin/gh/slayton58/46/orig 2025-12-04T09:20:54.4316784Z * [new branch] gh/slayton58/6/base -> origin/gh/slayton58/6/base 2025-12-04T09:20:54.4316945Z * [new branch] gh/slayton58/6/head -> origin/gh/slayton58/6/head 2025-12-04T09:20:54.4317090Z * [new branch] gh/slayton58/7/base -> origin/gh/slayton58/7/base 2025-12-04T09:20:54.4317232Z * [new branch] gh/slayton58/7/head -> origin/gh/slayton58/7/head 2025-12-04T09:20:54.4317401Z * [new branch] gh/soulitzer/269/base -> origin/gh/soulitzer/269/base 2025-12-04T09:20:54.4317553Z * [new branch] gh/soulitzer/269/head -> origin/gh/soulitzer/269/head 2025-12-04T09:20:54.4317717Z * [new branch] gh/soulitzer/269/orig -> origin/gh/soulitzer/269/orig 2025-12-04T09:20:54.4317867Z * [new branch] gh/soulitzer/276/base -> origin/gh/soulitzer/276/base 2025-12-04T09:20:54.4318020Z * [new branch] gh/soulitzer/276/head -> origin/gh/soulitzer/276/head 2025-12-04T09:20:54.4318177Z * [new branch] gh/soulitzer/276/orig -> origin/gh/soulitzer/276/orig 2025-12-04T09:20:54.4318726Z * [new branch] gh/soulitzer/287/base -> origin/gh/soulitzer/287/base 2025-12-04T09:20:54.4319443Z * [new branch] gh/soulitzer/287/head -> origin/gh/soulitzer/287/head 2025-12-04T09:20:54.4320033Z * [new branch] gh/soulitzer/287/orig -> origin/gh/soulitzer/287/orig 2025-12-04T09:20:54.4322370Z * [new branch] gh/soulitzer/296/base -> origin/gh/soulitzer/296/base 2025-12-04T09:20:54.4322938Z * [new branch] gh/soulitzer/296/head -> origin/gh/soulitzer/296/head 2025-12-04T09:20:54.4323133Z * [new branch] gh/soulitzer/296/orig -> origin/gh/soulitzer/296/orig 2025-12-04T09:20:54.4330335Z * [new branch] gh/soulitzer/299/base -> origin/gh/soulitzer/299/base 2025-12-04T09:20:54.4330633Z * [new branch] gh/soulitzer/299/head -> origin/gh/soulitzer/299/head 2025-12-04T09:20:54.4330812Z * [new branch] gh/soulitzer/299/orig -> origin/gh/soulitzer/299/orig 2025-12-04T09:20:54.4330973Z * [new branch] gh/soulitzer/300/base -> origin/gh/soulitzer/300/base 2025-12-04T09:20:54.4331132Z * [new branch] gh/soulitzer/300/head -> origin/gh/soulitzer/300/head 2025-12-04T09:20:54.4331291Z * [new branch] gh/soulitzer/300/orig -> origin/gh/soulitzer/300/orig 2025-12-04T09:20:54.4331439Z * [new branch] gh/soulitzer/301/base -> origin/gh/soulitzer/301/base 2025-12-04T09:20:54.4331591Z * [new branch] gh/soulitzer/301/head -> origin/gh/soulitzer/301/head 2025-12-04T09:20:54.4331993Z * [new branch] gh/soulitzer/301/orig -> origin/gh/soulitzer/301/orig 2025-12-04T09:20:54.4332885Z * [new branch] gh/soulitzer/313/base -> origin/gh/soulitzer/313/base 2025-12-04T09:20:54.4337569Z * [new branch] gh/soulitzer/313/head -> origin/gh/soulitzer/313/head 2025-12-04T09:20:54.4338315Z * [new branch] gh/soulitzer/313/orig -> origin/gh/soulitzer/313/orig 2025-12-04T09:20:54.4338482Z * [new branch] gh/soulitzer/319/base -> origin/gh/soulitzer/319/base 2025-12-04T09:20:54.4338649Z * [new branch] gh/soulitzer/319/head -> origin/gh/soulitzer/319/head 2025-12-04T09:20:54.4338802Z * [new branch] gh/soulitzer/319/orig -> origin/gh/soulitzer/319/orig 2025-12-04T09:20:54.4338961Z * [new branch] gh/soulitzer/320/base -> origin/gh/soulitzer/320/base 2025-12-04T09:20:54.4339123Z * [new branch] gh/soulitzer/320/head -> origin/gh/soulitzer/320/head 2025-12-04T09:20:54.4339295Z * [new branch] gh/soulitzer/320/orig -> origin/gh/soulitzer/320/orig 2025-12-04T09:20:54.4342490Z * [new branch] gh/soulitzer/336/base -> origin/gh/soulitzer/336/base 2025-12-04T09:20:54.4342810Z * [new branch] gh/soulitzer/336/head -> origin/gh/soulitzer/336/head 2025-12-04T09:20:54.4343009Z * [new branch] gh/soulitzer/336/orig -> origin/gh/soulitzer/336/orig 2025-12-04T09:20:54.4343265Z * [new branch] gh/soulitzer/347/base -> origin/gh/soulitzer/347/base 2025-12-04T09:20:54.4343990Z * [new branch] gh/soulitzer/347/head -> origin/gh/soulitzer/347/head 2025-12-04T09:20:54.4344731Z * [new branch] gh/soulitzer/347/orig -> origin/gh/soulitzer/347/orig 2025-12-04T09:20:54.4345701Z * [new branch] gh/soulitzer/349/base -> origin/gh/soulitzer/349/base 2025-12-04T09:20:54.4346023Z * [new branch] gh/soulitzer/349/head -> origin/gh/soulitzer/349/head 2025-12-04T09:20:54.4347465Z * [new branch] gh/soulitzer/349/orig -> origin/gh/soulitzer/349/orig 2025-12-04T09:20:54.4348291Z * [new branch] gh/soulitzer/350/base -> origin/gh/soulitzer/350/base 2025-12-04T09:20:54.4348758Z * [new branch] gh/soulitzer/350/head -> origin/gh/soulitzer/350/head 2025-12-04T09:20:54.4349731Z * [new branch] gh/soulitzer/350/orig -> origin/gh/soulitzer/350/orig 2025-12-04T09:20:54.4350837Z * [new branch] gh/soulitzer/351/base -> origin/gh/soulitzer/351/base 2025-12-04T09:20:54.4351231Z * [new branch] gh/soulitzer/351/head -> origin/gh/soulitzer/351/head 2025-12-04T09:20:54.4353018Z * [new branch] gh/soulitzer/351/orig -> origin/gh/soulitzer/351/orig 2025-12-04T09:20:54.4353232Z * [new branch] gh/soulitzer/353/base -> origin/gh/soulitzer/353/base 2025-12-04T09:20:54.4353743Z * [new branch] gh/soulitzer/353/head -> origin/gh/soulitzer/353/head 2025-12-04T09:20:54.4354722Z * [new branch] gh/soulitzer/353/orig -> origin/gh/soulitzer/353/orig 2025-12-04T09:20:54.4356249Z * [new branch] gh/soulitzer/358/base -> origin/gh/soulitzer/358/base 2025-12-04T09:20:54.4356770Z * [new branch] gh/soulitzer/358/head -> origin/gh/soulitzer/358/head 2025-12-04T09:20:54.4357842Z * [new branch] gh/soulitzer/358/orig -> origin/gh/soulitzer/358/orig 2025-12-04T09:20:54.4359974Z * [new branch] gh/soulitzer/359/base -> origin/gh/soulitzer/359/base 2025-12-04T09:20:54.4360182Z * [new branch] gh/soulitzer/359/head -> origin/gh/soulitzer/359/head 2025-12-04T09:20:54.4362365Z * [new branch] gh/soulitzer/359/orig -> origin/gh/soulitzer/359/orig 2025-12-04T09:20:54.4365315Z * [new branch] gh/soulitzer/374/base -> origin/gh/soulitzer/374/base 2025-12-04T09:20:54.4365634Z * [new branch] gh/soulitzer/374/head -> origin/gh/soulitzer/374/head 2025-12-04T09:20:54.4368084Z * [new branch] gh/soulitzer/374/orig -> origin/gh/soulitzer/374/orig 2025-12-04T09:20:54.4368251Z * [new branch] gh/soulitzer/375/base -> origin/gh/soulitzer/375/base 2025-12-04T09:20:54.4368902Z * [new branch] gh/soulitzer/375/head -> origin/gh/soulitzer/375/head 2025-12-04T09:20:54.4369094Z * [new branch] gh/soulitzer/375/orig -> origin/gh/soulitzer/375/orig 2025-12-04T09:20:54.4369259Z * [new branch] gh/soulitzer/380/base -> origin/gh/soulitzer/380/base 2025-12-04T09:20:54.4372067Z * [new branch] gh/soulitzer/380/head -> origin/gh/soulitzer/380/head 2025-12-04T09:20:54.4372227Z * [new branch] gh/soulitzer/380/orig -> origin/gh/soulitzer/380/orig 2025-12-04T09:20:54.4372394Z * [new branch] gh/soulitzer/385/base -> origin/gh/soulitzer/385/base 2025-12-04T09:20:54.4372544Z * [new branch] gh/soulitzer/385/head -> origin/gh/soulitzer/385/head 2025-12-04T09:20:54.4372695Z * [new branch] gh/soulitzer/385/orig -> origin/gh/soulitzer/385/orig 2025-12-04T09:20:54.4378656Z * [new branch] gh/soulitzer/386/base -> origin/gh/soulitzer/386/base 2025-12-04T09:20:54.4382918Z * [new branch] gh/soulitzer/386/head -> origin/gh/soulitzer/386/head 2025-12-04T09:20:54.4387140Z * [new branch] gh/soulitzer/386/orig -> origin/gh/soulitzer/386/orig 2025-12-04T09:20:54.4390721Z * [new branch] gh/soulitzer/387/base -> origin/gh/soulitzer/387/base 2025-12-04T09:20:54.4392949Z * [new branch] gh/soulitzer/387/head -> origin/gh/soulitzer/387/head 2025-12-04T09:20:54.4393113Z * [new branch] gh/soulitzer/387/orig -> origin/gh/soulitzer/387/orig 2025-12-04T09:20:54.4393306Z * [new branch] gh/soulitzer/388/base -> origin/gh/soulitzer/388/base 2025-12-04T09:20:54.4393450Z * [new branch] gh/soulitzer/388/head -> origin/gh/soulitzer/388/head 2025-12-04T09:20:54.4393602Z * [new branch] gh/soulitzer/388/orig -> origin/gh/soulitzer/388/orig 2025-12-04T09:20:54.4393755Z * [new branch] gh/soulitzer/389/base -> origin/gh/soulitzer/389/base 2025-12-04T09:20:54.4393899Z * [new branch] gh/soulitzer/389/head -> origin/gh/soulitzer/389/head 2025-12-04T09:20:54.4394052Z * [new branch] gh/soulitzer/389/orig -> origin/gh/soulitzer/389/orig 2025-12-04T09:20:54.4394195Z * [new branch] gh/soulitzer/390/base -> origin/gh/soulitzer/390/base 2025-12-04T09:20:54.4394347Z * [new branch] gh/soulitzer/390/head -> origin/gh/soulitzer/390/head 2025-12-04T09:20:54.4394492Z * [new branch] gh/soulitzer/390/orig -> origin/gh/soulitzer/390/orig 2025-12-04T09:20:54.4394642Z * [new branch] gh/soulitzer/391/base -> origin/gh/soulitzer/391/base 2025-12-04T09:20:54.4394793Z * [new branch] gh/soulitzer/391/head -> origin/gh/soulitzer/391/head 2025-12-04T09:20:54.4394937Z * [new branch] gh/soulitzer/391/orig -> origin/gh/soulitzer/391/orig 2025-12-04T09:20:54.4395093Z * [new branch] gh/soulitzer/392/base -> origin/gh/soulitzer/392/base 2025-12-04T09:20:54.4395247Z * [new branch] gh/soulitzer/392/head -> origin/gh/soulitzer/392/head 2025-12-04T09:20:54.4395390Z * [new branch] gh/soulitzer/392/orig -> origin/gh/soulitzer/392/orig 2025-12-04T09:20:54.4395549Z * [new branch] gh/swolchok/728/next -> origin/gh/swolchok/728/next 2025-12-04T09:20:54.4395696Z * [new branch] gh/swolchok/819/base -> origin/gh/swolchok/819/base 2025-12-04T09:20:54.4396073Z * [new branch] gh/swolchok/819/head -> origin/gh/swolchok/819/head 2025-12-04T09:20:54.4396254Z * [new branch] gh/swolchok/819/orig -> origin/gh/swolchok/819/orig 2025-12-04T09:20:54.4396398Z * [new branch] gh/swolchok/824/base -> origin/gh/swolchok/824/base 2025-12-04T09:20:54.4396615Z * [new branch] gh/swolchok/824/head -> origin/gh/swolchok/824/head 2025-12-04T09:20:54.4396765Z * [new branch] gh/swolchok/824/orig -> origin/gh/swolchok/824/orig 2025-12-04T09:20:54.4396913Z * [new branch] gh/swolchok/829/base -> origin/gh/swolchok/829/base 2025-12-04T09:20:54.4397065Z * [new branch] gh/swolchok/829/head -> origin/gh/swolchok/829/head 2025-12-04T09:20:54.4399197Z * [new branch] gh/swolchok/829/orig -> origin/gh/swolchok/829/orig 2025-12-04T09:20:54.4399654Z * [new branch] gh/swolchok/839/base -> origin/gh/swolchok/839/base 2025-12-04T09:20:54.4400176Z * [new branch] gh/swolchok/839/head -> origin/gh/swolchok/839/head 2025-12-04T09:20:54.4400362Z * [new branch] gh/swolchok/839/orig -> origin/gh/swolchok/839/orig 2025-12-04T09:20:54.4400523Z * [new branch] gh/swolchok/841/base -> origin/gh/swolchok/841/base 2025-12-04T09:20:54.4410725Z * [new branch] gh/swolchok/841/head -> origin/gh/swolchok/841/head 2025-12-04T09:20:54.4410896Z * [new branch] gh/swolchok/841/orig -> origin/gh/swolchok/841/orig 2025-12-04T09:20:54.4411282Z * [new branch] gh/swolchok/842/base -> origin/gh/swolchok/842/base 2025-12-04T09:20:54.4411447Z * [new branch] gh/swolchok/842/head -> origin/gh/swolchok/842/head 2025-12-04T09:20:54.4411608Z * [new branch] gh/swolchok/842/orig -> origin/gh/swolchok/842/orig 2025-12-04T09:20:54.4411762Z * [new branch] gh/swolchok/845/base -> origin/gh/swolchok/845/base 2025-12-04T09:20:54.4411930Z * [new branch] gh/swolchok/845/head -> origin/gh/swolchok/845/head 2025-12-04T09:20:54.4412074Z * [new branch] gh/swolchok/845/orig -> origin/gh/swolchok/845/orig 2025-12-04T09:20:54.4412217Z * [new branch] gh/swolchok/848/base -> origin/gh/swolchok/848/base 2025-12-04T09:20:54.4412376Z * [new branch] gh/swolchok/848/head -> origin/gh/swolchok/848/head 2025-12-04T09:20:54.4412520Z * [new branch] gh/swolchok/848/orig -> origin/gh/swolchok/848/orig 2025-12-04T09:20:54.4412671Z * [new branch] gh/swolchok/856/base -> origin/gh/swolchok/856/base 2025-12-04T09:20:54.4412812Z * [new branch] gh/swolchok/856/head -> origin/gh/swolchok/856/head 2025-12-04T09:20:54.4412955Z * [new branch] gh/swolchok/856/orig -> origin/gh/swolchok/856/orig 2025-12-04T09:20:54.4413106Z * [new branch] gh/swolchok/860/base -> origin/gh/swolchok/860/base 2025-12-04T09:20:54.4413252Z * [new branch] gh/swolchok/860/head -> origin/gh/swolchok/860/head 2025-12-04T09:20:54.4413408Z * [new branch] gh/swolchok/860/orig -> origin/gh/swolchok/860/orig 2025-12-04T09:20:54.4413552Z * [new branch] gh/swolchok/861/base -> origin/gh/swolchok/861/base 2025-12-04T09:20:54.4413699Z * [new branch] gh/swolchok/861/head -> origin/gh/swolchok/861/head 2025-12-04T09:20:54.4414274Z * [new branch] gh/swolchok/861/orig -> origin/gh/swolchok/861/orig 2025-12-04T09:20:54.4414424Z * [new branch] gh/swolchok/862/base -> origin/gh/swolchok/862/base 2025-12-04T09:20:54.4414724Z * [new branch] gh/swolchok/862/head -> origin/gh/swolchok/862/head 2025-12-04T09:20:54.4415717Z * [new branch] gh/swolchok/862/orig -> origin/gh/swolchok/862/orig 2025-12-04T09:20:54.4416919Z * [new branch] gh/swolchok/863/base -> origin/gh/swolchok/863/base 2025-12-04T09:20:54.4417329Z * [new branch] gh/swolchok/863/head -> origin/gh/swolchok/863/head 2025-12-04T09:20:54.4418352Z * [new branch] gh/swolchok/863/orig -> origin/gh/swolchok/863/orig 2025-12-04T09:20:54.4423056Z * [new branch] gh/swolchok/864/base -> origin/gh/swolchok/864/base 2025-12-04T09:20:54.4423257Z * [new branch] gh/swolchok/864/head -> origin/gh/swolchok/864/head 2025-12-04T09:20:54.4423422Z * [new branch] gh/swolchok/864/orig -> origin/gh/swolchok/864/orig 2025-12-04T09:20:54.4423572Z * [new branch] gh/swolchok/865/base -> origin/gh/swolchok/865/base 2025-12-04T09:20:54.4423726Z * [new branch] gh/swolchok/865/head -> origin/gh/swolchok/865/head 2025-12-04T09:20:54.4423879Z * [new branch] gh/swolchok/865/orig -> origin/gh/swolchok/865/orig 2025-12-04T09:20:54.4427968Z * [new branch] gh/swolchok/866/base -> origin/gh/swolchok/866/base 2025-12-04T09:20:54.4428315Z * [new branch] gh/swolchok/866/head -> origin/gh/swolchok/866/head 2025-12-04T09:20:54.4428509Z * [new branch] gh/swolchok/866/orig -> origin/gh/swolchok/866/orig 2025-12-04T09:20:54.4428728Z * [new branch] gh/swolchok/867/base -> origin/gh/swolchok/867/base 2025-12-04T09:20:54.4428926Z * [new branch] gh/swolchok/867/head -> origin/gh/swolchok/867/head 2025-12-04T09:20:54.4429183Z * [new branch] gh/swolchok/867/orig -> origin/gh/swolchok/867/orig 2025-12-04T09:20:54.4433785Z * [new branch] gh/swolchok/868/base -> origin/gh/swolchok/868/base 2025-12-04T09:20:54.4434254Z * [new branch] gh/swolchok/868/head -> origin/gh/swolchok/868/head 2025-12-04T09:20:54.4434405Z * [new branch] gh/swolchok/868/orig -> origin/gh/swolchok/868/orig 2025-12-04T09:20:54.4434580Z * [new branch] gh/swolchok/869/base -> origin/gh/swolchok/869/base 2025-12-04T09:20:54.4434728Z * [new branch] gh/swolchok/869/head -> origin/gh/swolchok/869/head 2025-12-04T09:20:54.4434882Z * [new branch] gh/swolchok/869/orig -> origin/gh/swolchok/869/orig 2025-12-04T09:20:54.4435429Z * [new branch] gh/swolchok/870/base -> origin/gh/swolchok/870/base 2025-12-04T09:20:54.4436758Z * [new branch] gh/swolchok/870/head -> origin/gh/swolchok/870/head 2025-12-04T09:20:54.4441373Z * [new branch] gh/swolchok/870/orig -> origin/gh/swolchok/870/orig 2025-12-04T09:20:54.4441571Z * [new branch] gh/swolchok/871/base -> origin/gh/swolchok/871/base 2025-12-04T09:20:54.4441950Z * [new branch] gh/swolchok/871/head -> origin/gh/swolchok/871/head 2025-12-04T09:20:54.4442114Z * [new branch] gh/swolchok/871/orig -> origin/gh/swolchok/871/orig 2025-12-04T09:20:54.4442305Z * [new branch] gh/teja-rao/4/base -> origin/gh/teja-rao/4/base 2025-12-04T09:20:54.4442449Z * [new branch] gh/teja-rao/4/head -> origin/gh/teja-rao/4/head 2025-12-04T09:20:54.4451906Z * [new branch] gh/teja-rao/4/orig -> origin/gh/teja-rao/4/orig 2025-12-04T09:20:54.4456674Z * [new branch] gh/tianyu-l/2/base -> origin/gh/tianyu-l/2/base 2025-12-04T09:20:54.4461848Z * [new branch] gh/tianyu-l/2/head -> origin/gh/tianyu-l/2/head 2025-12-04T09:20:54.4462049Z * [new branch] gh/tianyu-l/2/orig -> origin/gh/tianyu-l/2/orig 2025-12-04T09:20:54.4462481Z * [new branch] gh/tianyu-l/3/base -> origin/gh/tianyu-l/3/base 2025-12-04T09:20:54.4462655Z * [new branch] gh/tianyu-l/3/orig -> origin/gh/tianyu-l/3/orig 2025-12-04T09:20:54.4463188Z * [new branch] gh/tianyu-l/4/base -> origin/gh/tianyu-l/4/base 2025-12-04T09:20:54.4463347Z * [new branch] gh/tianyu-l/4/head -> origin/gh/tianyu-l/4/head 2025-12-04T09:20:54.4463497Z * [new branch] gh/tianyu-l/4/orig -> origin/gh/tianyu-l/4/orig 2025-12-04T09:20:54.4463786Z * [new branch] gh/tugsbayasgalan/10/base -> origin/gh/tugsbayasgalan/10/base 2025-12-04T09:20:54.4463995Z * [new branch] gh/tugsbayasgalan/10/head -> origin/gh/tugsbayasgalan/10/head 2025-12-04T09:20:54.4464173Z * [new branch] gh/tugsbayasgalan/10/orig -> origin/gh/tugsbayasgalan/10/orig 2025-12-04T09:20:54.4464354Z * [new branch] gh/tugsbayasgalan/13/base -> origin/gh/tugsbayasgalan/13/base 2025-12-04T09:20:54.4464533Z * [new branch] gh/tugsbayasgalan/13/head -> origin/gh/tugsbayasgalan/13/head 2025-12-04T09:20:54.4464718Z * [new branch] gh/tugsbayasgalan/13/orig -> origin/gh/tugsbayasgalan/13/orig 2025-12-04T09:20:54.4464903Z * [new branch] gh/tugsbayasgalan/17/base -> origin/gh/tugsbayasgalan/17/base 2025-12-04T09:20:54.4465086Z * [new branch] gh/tugsbayasgalan/17/head -> origin/gh/tugsbayasgalan/17/head 2025-12-04T09:20:54.4465259Z * [new branch] gh/tugsbayasgalan/17/orig -> origin/gh/tugsbayasgalan/17/orig 2025-12-04T09:20:54.4465445Z * [new branch] gh/tugsbayasgalan/2/base -> origin/gh/tugsbayasgalan/2/base 2025-12-04T09:20:54.4465626Z * [new branch] gh/tugsbayasgalan/2/head -> origin/gh/tugsbayasgalan/2/head 2025-12-04T09:20:54.4465793Z * [new branch] gh/tugsbayasgalan/2/orig -> origin/gh/tugsbayasgalan/2/orig 2025-12-04T09:20:54.4466073Z * [new branch] gh/tugsbayasgalan/28/base -> origin/gh/tugsbayasgalan/28/base 2025-12-04T09:20:54.4466259Z * [new branch] gh/tugsbayasgalan/28/head -> origin/gh/tugsbayasgalan/28/head 2025-12-04T09:20:54.4466433Z * [new branch] gh/tugsbayasgalan/28/orig -> origin/gh/tugsbayasgalan/28/orig 2025-12-04T09:20:54.4466614Z * [new branch] gh/tugsbayasgalan/32/base -> origin/gh/tugsbayasgalan/32/base 2025-12-04T09:20:54.4466787Z * [new branch] gh/tugsbayasgalan/32/head -> origin/gh/tugsbayasgalan/32/head 2025-12-04T09:20:54.4467974Z * [new branch] gh/tugsbayasgalan/32/orig -> origin/gh/tugsbayasgalan/32/orig 2025-12-04T09:20:54.4468241Z * [new branch] gh/tugsbayasgalan/35/base -> origin/gh/tugsbayasgalan/35/base 2025-12-04T09:20:54.4468431Z * [new branch] gh/tugsbayasgalan/35/head -> origin/gh/tugsbayasgalan/35/head 2025-12-04T09:20:54.4468698Z * [new branch] gh/tugsbayasgalan/35/orig -> origin/gh/tugsbayasgalan/35/orig 2025-12-04T09:20:54.4468897Z * [new branch] gh/tugsbayasgalan/36/base -> origin/gh/tugsbayasgalan/36/base 2025-12-04T09:20:54.4469161Z * [new branch] gh/tugsbayasgalan/36/head -> origin/gh/tugsbayasgalan/36/head 2025-12-04T09:20:54.4469361Z * [new branch] gh/tugsbayasgalan/36/orig -> origin/gh/tugsbayasgalan/36/orig 2025-12-04T09:20:54.4470200Z * [new branch] gh/tugsbayasgalan/37/base -> origin/gh/tugsbayasgalan/37/base 2025-12-04T09:20:54.4471304Z * [new branch] gh/tugsbayasgalan/37/head -> origin/gh/tugsbayasgalan/37/head 2025-12-04T09:20:54.4471871Z * [new branch] gh/tugsbayasgalan/37/orig -> origin/gh/tugsbayasgalan/37/orig 2025-12-04T09:20:54.4475638Z * [new branch] gh/tugsbayasgalan/43/base -> origin/gh/tugsbayasgalan/43/base 2025-12-04T09:20:54.4475850Z * [new branch] gh/tugsbayasgalan/43/head -> origin/gh/tugsbayasgalan/43/head 2025-12-04T09:20:54.4476042Z * [new branch] gh/tugsbayasgalan/43/orig -> origin/gh/tugsbayasgalan/43/orig 2025-12-04T09:20:54.4476220Z * [new branch] gh/tugsbayasgalan/48/base -> origin/gh/tugsbayasgalan/48/base 2025-12-04T09:20:54.4476782Z * [new branch] gh/tugsbayasgalan/48/head -> origin/gh/tugsbayasgalan/48/head 2025-12-04T09:20:54.4477470Z * [new branch] gh/tugsbayasgalan/48/orig -> origin/gh/tugsbayasgalan/48/orig 2025-12-04T09:20:54.4478517Z * [new branch] gh/tugsbayasgalan/51/base -> origin/gh/tugsbayasgalan/51/base 2025-12-04T09:20:54.4479152Z * [new branch] gh/tugsbayasgalan/51/head -> origin/gh/tugsbayasgalan/51/head 2025-12-04T09:20:54.4479670Z * [new branch] gh/tugsbayasgalan/51/orig -> origin/gh/tugsbayasgalan/51/orig 2025-12-04T09:20:54.4481543Z * [new branch] gh/tugsbayasgalan/52/base -> origin/gh/tugsbayasgalan/52/base 2025-12-04T09:20:54.4482428Z * [new branch] gh/tugsbayasgalan/52/head -> origin/gh/tugsbayasgalan/52/head 2025-12-04T09:20:54.4482751Z * [new branch] gh/tugsbayasgalan/52/orig -> origin/gh/tugsbayasgalan/52/orig 2025-12-04T09:20:54.4483158Z * [new branch] gh/tugsbayasgalan/53/base -> origin/gh/tugsbayasgalan/53/base 2025-12-04T09:20:54.4484293Z * [new branch] gh/tugsbayasgalan/53/head -> origin/gh/tugsbayasgalan/53/head 2025-12-04T09:20:54.4484617Z * [new branch] gh/tugsbayasgalan/53/orig -> origin/gh/tugsbayasgalan/53/orig 2025-12-04T09:20:54.4486005Z * [new branch] gh/tugsbayasgalan/55/base -> origin/gh/tugsbayasgalan/55/base 2025-12-04T09:20:54.4486539Z * [new branch] gh/tugsbayasgalan/55/head -> origin/gh/tugsbayasgalan/55/head 2025-12-04T09:20:54.4488406Z * [new branch] gh/tugsbayasgalan/55/orig -> origin/gh/tugsbayasgalan/55/orig 2025-12-04T09:20:54.4488618Z * [new branch] gh/tugsbayasgalan/59/base -> origin/gh/tugsbayasgalan/59/base 2025-12-04T09:20:54.4493407Z * [new branch] gh/tugsbayasgalan/59/head -> origin/gh/tugsbayasgalan/59/head 2025-12-04T09:20:54.4493623Z * [new branch] gh/tugsbayasgalan/59/orig -> origin/gh/tugsbayasgalan/59/orig 2025-12-04T09:20:54.4493825Z * [new branch] gh/tugsbayasgalan/6/base -> origin/gh/tugsbayasgalan/6/base 2025-12-04T09:20:54.4493995Z * [new branch] gh/tugsbayasgalan/6/head -> origin/gh/tugsbayasgalan/6/head 2025-12-04T09:20:54.4494173Z * [new branch] gh/tugsbayasgalan/6/orig -> origin/gh/tugsbayasgalan/6/orig 2025-12-04T09:20:54.4494352Z * [new branch] gh/tugsbayasgalan/60/base -> origin/gh/tugsbayasgalan/60/base 2025-12-04T09:20:54.4494560Z * [new branch] gh/tugsbayasgalan/60/head -> origin/gh/tugsbayasgalan/60/head 2025-12-04T09:20:54.4495261Z * [new branch] gh/tugsbayasgalan/60/orig -> origin/gh/tugsbayasgalan/60/orig 2025-12-04T09:20:54.4497583Z * [new branch] gh/tugsbayasgalan/61/base -> origin/gh/tugsbayasgalan/61/base 2025-12-04T09:20:54.4497788Z * [new branch] gh/tugsbayasgalan/61/head -> origin/gh/tugsbayasgalan/61/head 2025-12-04T09:20:54.4498155Z * [new branch] gh/tugsbayasgalan/61/orig -> origin/gh/tugsbayasgalan/61/orig 2025-12-04T09:20:54.4500178Z * [new branch] gh/tugsbayasgalan/63/base -> origin/gh/tugsbayasgalan/63/base 2025-12-04T09:20:54.4500412Z * [new branch] gh/tugsbayasgalan/63/head -> origin/gh/tugsbayasgalan/63/head 2025-12-04T09:20:54.4501364Z * [new branch] gh/tugsbayasgalan/63/orig -> origin/gh/tugsbayasgalan/63/orig 2025-12-04T09:20:54.4501546Z * [new branch] gh/tugsbayasgalan/67/base -> origin/gh/tugsbayasgalan/67/base 2025-12-04T09:20:54.4502630Z * [new branch] gh/tugsbayasgalan/67/head -> origin/gh/tugsbayasgalan/67/head 2025-12-04T09:20:54.4503053Z * [new branch] gh/tugsbayasgalan/67/orig -> origin/gh/tugsbayasgalan/67/orig 2025-12-04T09:20:54.4504505Z * [new branch] gh/tugsbayasgalan/68/base -> origin/gh/tugsbayasgalan/68/base 2025-12-04T09:20:54.4505153Z * [new branch] gh/tugsbayasgalan/68/head -> origin/gh/tugsbayasgalan/68/head 2025-12-04T09:20:54.4505811Z * [new branch] gh/tugsbayasgalan/68/orig -> origin/gh/tugsbayasgalan/68/orig 2025-12-04T09:20:54.4507021Z * [new branch] gh/tugsbayasgalan/7/base -> origin/gh/tugsbayasgalan/7/base 2025-12-04T09:20:54.4507358Z * [new branch] gh/tugsbayasgalan/7/head -> origin/gh/tugsbayasgalan/7/head 2025-12-04T09:20:54.4508647Z * [new branch] gh/tugsbayasgalan/7/orig -> origin/gh/tugsbayasgalan/7/orig 2025-12-04T09:20:54.4510108Z * [new branch] gh/tugsbayasgalan/70/base -> origin/gh/tugsbayasgalan/70/base 2025-12-04T09:20:54.4510279Z * [new branch] gh/tugsbayasgalan/70/head -> origin/gh/tugsbayasgalan/70/head 2025-12-04T09:20:54.4511382Z * [new branch] gh/tugsbayasgalan/70/orig -> origin/gh/tugsbayasgalan/70/orig 2025-12-04T09:20:54.4512493Z * [new branch] gh/tugsbayasgalan/71/base -> origin/gh/tugsbayasgalan/71/base 2025-12-04T09:20:54.4513133Z * [new branch] gh/tugsbayasgalan/71/head -> origin/gh/tugsbayasgalan/71/head 2025-12-04T09:20:54.4514108Z * [new branch] gh/tugsbayasgalan/71/orig -> origin/gh/tugsbayasgalan/71/orig 2025-12-04T09:20:54.4515775Z * [new branch] gh/tugsbayasgalan/72/base -> origin/gh/tugsbayasgalan/72/base 2025-12-04T09:20:54.4515962Z * [new branch] gh/tugsbayasgalan/72/head -> origin/gh/tugsbayasgalan/72/head 2025-12-04T09:20:54.4518138Z * [new branch] gh/tugsbayasgalan/72/orig -> origin/gh/tugsbayasgalan/72/orig 2025-12-04T09:20:54.4518391Z * [new branch] gh/tugsbayasgalan/73/base -> origin/gh/tugsbayasgalan/73/base 2025-12-04T09:20:54.4518574Z * [new branch] gh/tugsbayasgalan/73/head -> origin/gh/tugsbayasgalan/73/head 2025-12-04T09:20:54.4519188Z * [new branch] gh/tugsbayasgalan/73/orig -> origin/gh/tugsbayasgalan/73/orig 2025-12-04T09:20:54.4522352Z * [new branch] gh/tugsbayasgalan/74/base -> origin/gh/tugsbayasgalan/74/base 2025-12-04T09:20:54.4522565Z * [new branch] gh/tugsbayasgalan/74/head -> origin/gh/tugsbayasgalan/74/head 2025-12-04T09:20:54.4522732Z * [new branch] gh/tugsbayasgalan/74/orig -> origin/gh/tugsbayasgalan/74/orig 2025-12-04T09:20:54.4523075Z * [new branch] gh/tugsbayasgalan/75/base -> origin/gh/tugsbayasgalan/75/base 2025-12-04T09:20:54.4523713Z * [new branch] gh/tugsbayasgalan/75/head -> origin/gh/tugsbayasgalan/75/head 2025-12-04T09:20:54.4524400Z * [new branch] gh/tugsbayasgalan/75/orig -> origin/gh/tugsbayasgalan/75/orig 2025-12-04T09:20:54.4527240Z * [new branch] gh/tugsbayasgalan/76/base -> origin/gh/tugsbayasgalan/76/base 2025-12-04T09:20:54.4527635Z * [new branch] gh/tugsbayasgalan/76/head -> origin/gh/tugsbayasgalan/76/head 2025-12-04T09:20:54.4527825Z * [new branch] gh/tugsbayasgalan/76/orig -> origin/gh/tugsbayasgalan/76/orig 2025-12-04T09:20:54.4528060Z * [new branch] gh/tugsbayasgalan/77/base -> origin/gh/tugsbayasgalan/77/base 2025-12-04T09:20:54.4528557Z * [new branch] gh/tugsbayasgalan/77/head -> origin/gh/tugsbayasgalan/77/head 2025-12-04T09:20:54.4529599Z * [new branch] gh/tugsbayasgalan/77/orig -> origin/gh/tugsbayasgalan/77/orig 2025-12-04T09:20:54.4535224Z * [new branch] gh/tugsbayasgalan/78/base -> origin/gh/tugsbayasgalan/78/base 2025-12-04T09:20:54.4540973Z * [new branch] gh/tugsbayasgalan/78/head -> origin/gh/tugsbayasgalan/78/head 2025-12-04T09:20:54.4542960Z * [new branch] gh/tugsbayasgalan/78/orig -> origin/gh/tugsbayasgalan/78/orig 2025-12-04T09:20:54.4543530Z * [new branch] gh/tugsbayasgalan/79/base -> origin/gh/tugsbayasgalan/79/base 2025-12-04T09:20:54.4543754Z * [new branch] gh/tugsbayasgalan/79/head -> origin/gh/tugsbayasgalan/79/head 2025-12-04T09:20:54.4544176Z * [new branch] gh/tugsbayasgalan/79/orig -> origin/gh/tugsbayasgalan/79/orig 2025-12-04T09:20:54.4544366Z * [new branch] gh/tugsbayasgalan/8/base -> origin/gh/tugsbayasgalan/8/base 2025-12-04T09:20:54.4544545Z * [new branch] gh/tugsbayasgalan/8/head -> origin/gh/tugsbayasgalan/8/head 2025-12-04T09:20:54.4545005Z * [new branch] gh/tugsbayasgalan/8/orig -> origin/gh/tugsbayasgalan/8/orig 2025-12-04T09:20:54.4545185Z * [new branch] gh/tugsbayasgalan/80/base -> origin/gh/tugsbayasgalan/80/base 2025-12-04T09:20:54.4545365Z * [new branch] gh/tugsbayasgalan/80/head -> origin/gh/tugsbayasgalan/80/head 2025-12-04T09:20:54.4547237Z * [new branch] gh/tugsbayasgalan/80/orig -> origin/gh/tugsbayasgalan/80/orig 2025-12-04T09:20:54.4547442Z * [new branch] gh/tugsbayasgalan/81/base -> origin/gh/tugsbayasgalan/81/base 2025-12-04T09:20:54.4547649Z * [new branch] gh/tugsbayasgalan/81/head -> origin/gh/tugsbayasgalan/81/head 2025-12-04T09:20:54.4547840Z * [new branch] gh/tugsbayasgalan/81/orig -> origin/gh/tugsbayasgalan/81/orig 2025-12-04T09:20:54.4556090Z * [new branch] gh/tugsbayasgalan/82/base -> origin/gh/tugsbayasgalan/82/base 2025-12-04T09:20:54.4560440Z * [new branch] gh/tugsbayasgalan/82/head -> origin/gh/tugsbayasgalan/82/head 2025-12-04T09:20:54.4560679Z * [new branch] gh/tugsbayasgalan/82/orig -> origin/gh/tugsbayasgalan/82/orig 2025-12-04T09:20:54.4560892Z * [new branch] gh/tugsbayasgalan/83/base -> origin/gh/tugsbayasgalan/83/base 2025-12-04T09:20:54.4561076Z * [new branch] gh/tugsbayasgalan/83/head -> origin/gh/tugsbayasgalan/83/head 2025-12-04T09:20:54.4561251Z * [new branch] gh/tugsbayasgalan/83/orig -> origin/gh/tugsbayasgalan/83/orig 2025-12-04T09:20:54.4561428Z * [new branch] gh/tugsbayasgalan/84/base -> origin/gh/tugsbayasgalan/84/base 2025-12-04T09:20:54.4561604Z * [new branch] gh/tugsbayasgalan/84/head -> origin/gh/tugsbayasgalan/84/head 2025-12-04T09:20:54.4561937Z * [new branch] gh/tugsbayasgalan/84/orig -> origin/gh/tugsbayasgalan/84/orig 2025-12-04T09:20:54.4562117Z * [new branch] gh/tugsbayasgalan/85/base -> origin/gh/tugsbayasgalan/85/base 2025-12-04T09:20:54.4562287Z * [new branch] gh/tugsbayasgalan/85/head -> origin/gh/tugsbayasgalan/85/head 2025-12-04T09:20:54.4562453Z * [new branch] gh/tugsbayasgalan/85/orig -> origin/gh/tugsbayasgalan/85/orig 2025-12-04T09:20:54.4562624Z * [new branch] gh/tugsbayasgalan/86/base -> origin/gh/tugsbayasgalan/86/base 2025-12-04T09:20:54.4562785Z * [new branch] gh/tugsbayasgalan/86/head -> origin/gh/tugsbayasgalan/86/head 2025-12-04T09:20:54.4562947Z * [new branch] gh/tugsbayasgalan/86/orig -> origin/gh/tugsbayasgalan/86/orig 2025-12-04T09:20:54.4563121Z * [new branch] gh/tugsbayasgalan/87/base -> origin/gh/tugsbayasgalan/87/base 2025-12-04T09:20:54.4563286Z * [new branch] gh/tugsbayasgalan/87/head -> origin/gh/tugsbayasgalan/87/head 2025-12-04T09:20:54.4567839Z * [new branch] gh/tugsbayasgalan/87/orig -> origin/gh/tugsbayasgalan/87/orig 2025-12-04T09:20:54.4572805Z * [new branch] gh/tugsbayasgalan/88/base -> origin/gh/tugsbayasgalan/88/base 2025-12-04T09:20:54.4577085Z * [new branch] gh/tugsbayasgalan/88/head -> origin/gh/tugsbayasgalan/88/head 2025-12-04T09:20:54.4577626Z * [new branch] gh/tugsbayasgalan/88/orig -> origin/gh/tugsbayasgalan/88/orig 2025-12-04T09:20:54.4577836Z * [new branch] gh/tugsbayasgalan/89/base -> origin/gh/tugsbayasgalan/89/base 2025-12-04T09:20:54.4578025Z * [new branch] gh/tugsbayasgalan/89/head -> origin/gh/tugsbayasgalan/89/head 2025-12-04T09:20:54.4578203Z * [new branch] gh/tugsbayasgalan/89/orig -> origin/gh/tugsbayasgalan/89/orig 2025-12-04T09:20:54.4578583Z * [new branch] gh/tugsbayasgalan/9/base -> origin/gh/tugsbayasgalan/9/base 2025-12-04T09:20:54.4578762Z * [new branch] gh/tugsbayasgalan/9/head -> origin/gh/tugsbayasgalan/9/head 2025-12-04T09:20:54.4578971Z * [new branch] gh/tugsbayasgalan/9/orig -> origin/gh/tugsbayasgalan/9/orig 2025-12-04T09:20:54.4579195Z * [new branch] gh/tugsbayasgalan/90/base -> origin/gh/tugsbayasgalan/90/base 2025-12-04T09:20:54.4579378Z * [new branch] gh/tugsbayasgalan/90/head -> origin/gh/tugsbayasgalan/90/head 2025-12-04T09:20:54.4579551Z * [new branch] gh/tugsbayasgalan/90/orig -> origin/gh/tugsbayasgalan/90/orig 2025-12-04T09:20:54.4579721Z * [new branch] gh/tugsbayasgalan/91/base -> origin/gh/tugsbayasgalan/91/base 2025-12-04T09:20:54.4579901Z * [new branch] gh/tugsbayasgalan/91/head -> origin/gh/tugsbayasgalan/91/head 2025-12-04T09:20:54.4580077Z * [new branch] gh/tugsbayasgalan/91/orig -> origin/gh/tugsbayasgalan/91/orig 2025-12-04T09:20:54.4580252Z * [new branch] gh/tugsbayasgalan/92/base -> origin/gh/tugsbayasgalan/92/base 2025-12-04T09:20:54.4580426Z * [new branch] gh/tugsbayasgalan/92/head -> origin/gh/tugsbayasgalan/92/head 2025-12-04T09:20:54.4580641Z * [new branch] gh/tugsbayasgalan/92/orig -> origin/gh/tugsbayasgalan/92/orig 2025-12-04T09:20:54.4580822Z * [new branch] gh/tugsbayasgalan/93/base -> origin/gh/tugsbayasgalan/93/base 2025-12-04T09:20:54.4581000Z * [new branch] gh/tugsbayasgalan/93/head -> origin/gh/tugsbayasgalan/93/head 2025-12-04T09:20:54.4581174Z * [new branch] gh/tugsbayasgalan/93/orig -> origin/gh/tugsbayasgalan/93/orig 2025-12-04T09:20:54.4582094Z * [new branch] gh/v0i0/14/base -> origin/gh/v0i0/14/base 2025-12-04T09:20:54.4583288Z * [new branch] gh/v0i0/14/head -> origin/gh/v0i0/14/head 2025-12-04T09:20:54.4583679Z * [new branch] gh/v0i0/14/orig -> origin/gh/v0i0/14/orig 2025-12-04T09:20:54.4584806Z * [new branch] gh/v0i0/15/base -> origin/gh/v0i0/15/base 2025-12-04T09:20:54.4585349Z * [new branch] gh/v0i0/15/head -> origin/gh/v0i0/15/head 2025-12-04T09:20:54.4586353Z * [new branch] gh/v0i0/15/orig -> origin/gh/v0i0/15/orig 2025-12-04T09:20:54.4587390Z * [new branch] gh/v0i0/16/base -> origin/gh/v0i0/16/base 2025-12-04T09:20:54.4587852Z * [new branch] gh/v0i0/16/head -> origin/gh/v0i0/16/head 2025-12-04T09:20:54.4588919Z * [new branch] gh/v0i0/16/orig -> origin/gh/v0i0/16/orig 2025-12-04T09:20:54.4589899Z * [new branch] gh/v0i0/17/base -> origin/gh/v0i0/17/base 2025-12-04T09:20:54.4590625Z * [new branch] gh/v0i0/17/head -> origin/gh/v0i0/17/head 2025-12-04T09:20:54.4591212Z * [new branch] gh/v0i0/17/orig -> origin/gh/v0i0/17/orig 2025-12-04T09:20:54.4592542Z * [new branch] gh/v0i0/18/base -> origin/gh/v0i0/18/base 2025-12-04T09:20:54.4593049Z * [new branch] gh/v0i0/18/head -> origin/gh/v0i0/18/head 2025-12-04T09:20:54.4594070Z * [new branch] gh/v0i0/18/orig -> origin/gh/v0i0/18/orig 2025-12-04T09:20:54.4594943Z * [new branch] gh/v0i0/19/base -> origin/gh/v0i0/19/base 2025-12-04T09:20:54.4595875Z * [new branch] gh/v0i0/19/head -> origin/gh/v0i0/19/head 2025-12-04T09:20:54.4597145Z * [new branch] gh/v0i0/19/orig -> origin/gh/v0i0/19/orig 2025-12-04T09:20:54.4597705Z * [new branch] gh/vishal9-team/1/base -> origin/gh/vishal9-team/1/base 2025-12-04T09:20:54.4598634Z * [new branch] gh/vishal9-team/1/head -> origin/gh/vishal9-team/1/head 2025-12-04T09:20:54.4599371Z * [new branch] gh/vishal9-team/2/base -> origin/gh/vishal9-team/2/base 2025-12-04T09:20:54.4600296Z * [new branch] gh/vishal9-team/2/head -> origin/gh/vishal9-team/2/head 2025-12-04T09:20:54.4600809Z * [new branch] gh/vishal9-team/2/orig -> origin/gh/vishal9-team/2/orig 2025-12-04T09:20:54.4602453Z * [new branch] gh/vishal9-team/3/base -> origin/gh/vishal9-team/3/base 2025-12-04T09:20:54.4603905Z * [new branch] gh/vishal9-team/3/head -> origin/gh/vishal9-team/3/head 2025-12-04T09:20:54.4604507Z * [new branch] gh/vishal9-team/3/orig -> origin/gh/vishal9-team/3/orig 2025-12-04T09:20:54.4604705Z * [new branch] gh/vishal9-team/4/base -> origin/gh/vishal9-team/4/base 2025-12-04T09:20:54.4605192Z * [new branch] gh/vishal9-team/4/head -> origin/gh/vishal9-team/4/head 2025-12-04T09:20:54.4605973Z * [new branch] gh/vishal9-team/4/orig -> origin/gh/vishal9-team/4/orig 2025-12-04T09:20:54.4607365Z * [new branch] gh/vkuzo/1/next -> origin/gh/vkuzo/1/next 2025-12-04T09:20:54.4608003Z * [new branch] gh/vkuzo/2/next -> origin/gh/vkuzo/2/next 2025-12-04T09:20:54.4609455Z * [new branch] gh/vkuzo/3/next -> origin/gh/vkuzo/3/next 2025-12-04T09:20:54.4610965Z * [new branch] gh/wconstab/424/base -> origin/gh/wconstab/424/base 2025-12-04T09:20:54.4615517Z * [new branch] gh/wconstab/424/head -> origin/gh/wconstab/424/head 2025-12-04T09:20:54.4616142Z * [new branch] gh/wconstab/424/orig -> origin/gh/wconstab/424/orig 2025-12-04T09:20:54.4616334Z * [new branch] gh/wconstab/435/base -> origin/gh/wconstab/435/base 2025-12-04T09:20:54.4616490Z * [new branch] gh/wconstab/435/head -> origin/gh/wconstab/435/head 2025-12-04T09:20:54.4616640Z * [new branch] gh/wconstab/435/orig -> origin/gh/wconstab/435/orig 2025-12-04T09:20:54.4616833Z * [new branch] gh/wconstab/444/base -> origin/gh/wconstab/444/base 2025-12-04T09:20:54.4616977Z * [new branch] gh/wconstab/444/head -> origin/gh/wconstab/444/head 2025-12-04T09:20:54.4622592Z * [new branch] gh/wconstab/444/orig -> origin/gh/wconstab/444/orig 2025-12-04T09:20:54.4626990Z * [new branch] gh/wconstab/447/base -> origin/gh/wconstab/447/base 2025-12-04T09:20:54.4628666Z * [new branch] gh/wconstab/447/head -> origin/gh/wconstab/447/head 2025-12-04T09:20:54.4628840Z * [new branch] gh/wconstab/447/orig -> origin/gh/wconstab/447/orig 2025-12-04T09:20:54.4628987Z * [new branch] gh/wconstab/448/base -> origin/gh/wconstab/448/base 2025-12-04T09:20:54.4629141Z * [new branch] gh/wconstab/448/head -> origin/gh/wconstab/448/head 2025-12-04T09:20:54.4629297Z * [new branch] gh/wconstab/448/orig -> origin/gh/wconstab/448/orig 2025-12-04T09:20:54.4629443Z * [new branch] gh/wconstab/449/base -> origin/gh/wconstab/449/base 2025-12-04T09:20:54.4629597Z * [new branch] gh/wconstab/449/head -> origin/gh/wconstab/449/head 2025-12-04T09:20:54.4629747Z * [new branch] gh/wconstab/449/orig -> origin/gh/wconstab/449/orig 2025-12-04T09:20:54.4629896Z * [new branch] gh/wconstab/450/base -> origin/gh/wconstab/450/base 2025-12-04T09:20:54.4630038Z * [new branch] gh/wconstab/450/head -> origin/gh/wconstab/450/head 2025-12-04T09:20:54.4630335Z * [new branch] gh/wconstab/450/orig -> origin/gh/wconstab/450/orig 2025-12-04T09:20:54.4630559Z * [new branch] gh/wconstab/451/base -> origin/gh/wconstab/451/base 2025-12-04T09:20:54.4630707Z * [new branch] gh/wconstab/451/head -> origin/gh/wconstab/451/head 2025-12-04T09:20:54.4631086Z * [new branch] gh/wconstab/451/orig -> origin/gh/wconstab/451/orig 2025-12-04T09:20:54.4631236Z * [new branch] gh/wconstab/452/base -> origin/gh/wconstab/452/base 2025-12-04T09:20:54.4631386Z * [new branch] gh/wconstab/452/head -> origin/gh/wconstab/452/head 2025-12-04T09:20:54.4631869Z * [new branch] gh/wconstab/452/orig -> origin/gh/wconstab/452/orig 2025-12-04T09:20:54.4632675Z * [new branch] gh/wconstab/453/base -> origin/gh/wconstab/453/base 2025-12-04T09:20:54.4633787Z * [new branch] gh/wconstab/453/head -> origin/gh/wconstab/453/head 2025-12-04T09:20:54.4634229Z * [new branch] gh/wconstab/453/orig -> origin/gh/wconstab/453/orig 2025-12-04T09:20:54.4635445Z * [new branch] gh/wconstab/454/base -> origin/gh/wconstab/454/base 2025-12-04T09:20:54.4637738Z * [new branch] gh/wconstab/454/head -> origin/gh/wconstab/454/head 2025-12-04T09:20:54.4641033Z * [new branch] gh/wconstab/454/orig -> origin/gh/wconstab/454/orig 2025-12-04T09:20:54.4641210Z * [new branch] gh/wconstab/455/base -> origin/gh/wconstab/455/base 2025-12-04T09:20:54.4641360Z * [new branch] gh/wconstab/455/head -> origin/gh/wconstab/455/head 2025-12-04T09:20:54.4641541Z * [new branch] gh/wconstab/455/orig -> origin/gh/wconstab/455/orig 2025-12-04T09:20:54.4641758Z * [new branch] gh/wconstab/456/base -> origin/gh/wconstab/456/base 2025-12-04T09:20:54.4641920Z * [new branch] gh/wconstab/456/head -> origin/gh/wconstab/456/head 2025-12-04T09:20:54.4642296Z * [new branch] gh/wconstab/456/orig -> origin/gh/wconstab/456/orig 2025-12-04T09:20:54.4648283Z * [new branch] gh/wconstab/457/base -> origin/gh/wconstab/457/base 2025-12-04T09:20:54.4650233Z * [new branch] gh/wconstab/457/head -> origin/gh/wconstab/457/head 2025-12-04T09:20:54.4650413Z * [new branch] gh/wconstab/457/orig -> origin/gh/wconstab/457/orig 2025-12-04T09:20:54.4650574Z * [new branch] gh/wconstab/458/base -> origin/gh/wconstab/458/base 2025-12-04T09:20:54.4650731Z * [new branch] gh/wconstab/458/head -> origin/gh/wconstab/458/head 2025-12-04T09:20:54.4650877Z * [new branch] gh/wconstab/458/orig -> origin/gh/wconstab/458/orig 2025-12-04T09:20:54.4651032Z * [new branch] gh/wconstab/459/base -> origin/gh/wconstab/459/base 2025-12-04T09:20:54.4651178Z * [new branch] gh/wconstab/459/head -> origin/gh/wconstab/459/head 2025-12-04T09:20:54.4651327Z * [new branch] gh/wconstab/459/orig -> origin/gh/wconstab/459/orig 2025-12-04T09:20:54.4651618Z * [new branch] gh/wconstab/460/base -> origin/gh/wconstab/460/base 2025-12-04T09:20:54.4652117Z * [new branch] gh/wconstab/460/head -> origin/gh/wconstab/460/head 2025-12-04T09:20:54.4652845Z * [new branch] gh/wconstab/460/orig -> origin/gh/wconstab/460/orig 2025-12-04T09:20:54.4655341Z * [new branch] gh/wconstab/461/base -> origin/gh/wconstab/461/base 2025-12-04T09:20:54.4655542Z * [new branch] gh/wconstab/461/head -> origin/gh/wconstab/461/head 2025-12-04T09:20:54.4655694Z * [new branch] gh/wconstab/461/orig -> origin/gh/wconstab/461/orig 2025-12-04T09:20:54.4657805Z * [new branch] gh/wconstab/462/base -> origin/gh/wconstab/462/base 2025-12-04T09:20:54.4658002Z * [new branch] gh/wconstab/462/head -> origin/gh/wconstab/462/head 2025-12-04T09:20:54.4658164Z * [new branch] gh/wconstab/462/orig -> origin/gh/wconstab/462/orig 2025-12-04T09:20:54.4664123Z * [new branch] gh/wconstab/463/base -> origin/gh/wconstab/463/base 2025-12-04T09:20:54.4664505Z * [new branch] gh/wconstab/463/head -> origin/gh/wconstab/463/head 2025-12-04T09:20:54.4664671Z * [new branch] gh/wconstab/463/orig -> origin/gh/wconstab/463/orig 2025-12-04T09:20:54.4664817Z * [new branch] gh/wconstab/464/base -> origin/gh/wconstab/464/base 2025-12-04T09:20:54.4665033Z * [new branch] gh/wconstab/464/head -> origin/gh/wconstab/464/head 2025-12-04T09:20:54.4665197Z * [new branch] gh/wconstab/464/orig -> origin/gh/wconstab/464/orig 2025-12-04T09:20:54.4665347Z * [new branch] gh/wconstab/465/base -> origin/gh/wconstab/465/base 2025-12-04T09:20:54.4665500Z * [new branch] gh/wconstab/465/head -> origin/gh/wconstab/465/head 2025-12-04T09:20:54.4665919Z * [new branch] gh/wconstab/465/orig -> origin/gh/wconstab/465/orig 2025-12-04T09:20:54.4666217Z * [new branch] gh/wconstab/466/base -> origin/gh/wconstab/466/base 2025-12-04T09:20:54.4667849Z * [new branch] gh/wconstab/466/head -> origin/gh/wconstab/466/head 2025-12-04T09:20:54.4668014Z * [new branch] gh/wconstab/466/orig -> origin/gh/wconstab/466/orig 2025-12-04T09:20:54.4674410Z * [new branch] gh/wconstab/467/base -> origin/gh/wconstab/467/base 2025-12-04T09:20:54.4674750Z * [new branch] gh/wconstab/467/head -> origin/gh/wconstab/467/head 2025-12-04T09:20:54.4675087Z * [new branch] gh/wconstab/467/orig -> origin/gh/wconstab/467/orig 2025-12-04T09:20:54.4675246Z * [new branch] gh/wconstab/468/base -> origin/gh/wconstab/468/base 2025-12-04T09:20:54.4675402Z * [new branch] gh/wconstab/468/head -> origin/gh/wconstab/468/head 2025-12-04T09:20:54.4675557Z * [new branch] gh/wconstab/468/orig -> origin/gh/wconstab/468/orig 2025-12-04T09:20:54.4675858Z * [new branch] gh/weifengpy/39/base -> origin/gh/weifengpy/39/base 2025-12-04T09:20:54.4676018Z * [new branch] gh/weifengpy/39/head -> origin/gh/weifengpy/39/head 2025-12-04T09:20:54.4676319Z * [new branch] gh/weifengpy/39/orig -> origin/gh/weifengpy/39/orig 2025-12-04T09:20:54.4676568Z * [new branch] gh/weifengpy/40/base -> origin/gh/weifengpy/40/base 2025-12-04T09:20:54.4682800Z * [new branch] gh/weifengpy/40/head -> origin/gh/weifengpy/40/head 2025-12-04T09:20:54.4682981Z * [new branch] gh/weifengpy/40/orig -> origin/gh/weifengpy/40/orig 2025-12-04T09:20:54.4687841Z * [new branch] gh/weifengpy/41/base -> origin/gh/weifengpy/41/base 2025-12-04T09:20:54.4692137Z * [new branch] gh/weifengpy/41/head -> origin/gh/weifengpy/41/head 2025-12-04T09:20:54.4696778Z * [new branch] gh/weifengpy/41/orig -> origin/gh/weifengpy/41/orig 2025-12-04T09:20:54.4701929Z * [new branch] gh/williamwen42/250/base -> origin/gh/williamwen42/250/base 2025-12-04T09:20:54.4707777Z * [new branch] gh/williamwen42/250/head -> origin/gh/williamwen42/250/head 2025-12-04T09:20:54.4708266Z * [new branch] gh/williamwen42/250/orig -> origin/gh/williamwen42/250/orig 2025-12-04T09:20:54.4708489Z * [new branch] gh/williamwen42/279/base -> origin/gh/williamwen42/279/base 2025-12-04T09:20:54.4708646Z * [new branch] gh/williamwen42/279/head -> origin/gh/williamwen42/279/head 2025-12-04T09:20:54.4708801Z * [new branch] gh/williamwen42/279/orig -> origin/gh/williamwen42/279/orig 2025-12-04T09:20:54.4708961Z * [new branch] gh/williamwen42/282/base -> origin/gh/williamwen42/282/base 2025-12-04T09:20:54.4709113Z * [new branch] gh/williamwen42/282/head -> origin/gh/williamwen42/282/head 2025-12-04T09:20:54.4709264Z * [new branch] gh/williamwen42/282/orig -> origin/gh/williamwen42/282/orig 2025-12-04T09:20:54.4709587Z * [new branch] gh/williamwen42/287/base -> origin/gh/williamwen42/287/base 2025-12-04T09:20:54.4709738Z * [new branch] gh/williamwen42/287/head -> origin/gh/williamwen42/287/head 2025-12-04T09:20:54.4709900Z * [new branch] gh/williamwen42/287/orig -> origin/gh/williamwen42/287/orig 2025-12-04T09:20:54.4710101Z * [new branch] gh/williamwen42/288/base -> origin/gh/williamwen42/288/base 2025-12-04T09:20:54.4710252Z * [new branch] gh/williamwen42/288/head -> origin/gh/williamwen42/288/head 2025-12-04T09:20:54.4710432Z * [new branch] gh/williamwen42/288/orig -> origin/gh/williamwen42/288/orig 2025-12-04T09:20:54.4710612Z * [new branch] gh/williamwen42/296/base -> origin/gh/williamwen42/296/base 2025-12-04T09:20:54.4710772Z * [new branch] gh/williamwen42/296/head -> origin/gh/williamwen42/296/head 2025-12-04T09:20:54.4710926Z * [new branch] gh/williamwen42/296/orig -> origin/gh/williamwen42/296/orig 2025-12-04T09:20:54.4711075Z * [new branch] gh/williamwen42/297/base -> origin/gh/williamwen42/297/base 2025-12-04T09:20:54.4711232Z * [new branch] gh/williamwen42/297/head -> origin/gh/williamwen42/297/head 2025-12-04T09:20:54.4711386Z * [new branch] gh/williamwen42/297/orig -> origin/gh/williamwen42/297/orig 2025-12-04T09:20:54.4711545Z * [new branch] gh/williamwen42/306/base -> origin/gh/williamwen42/306/base 2025-12-04T09:20:54.4711696Z * [new branch] gh/williamwen42/306/head -> origin/gh/williamwen42/306/head 2025-12-04T09:20:54.4711845Z * [new branch] gh/williamwen42/306/orig -> origin/gh/williamwen42/306/orig 2025-12-04T09:20:54.4712004Z * [new branch] gh/williamwen42/309/base -> origin/gh/williamwen42/309/base 2025-12-04T09:20:54.4712157Z * [new branch] gh/williamwen42/309/head -> origin/gh/williamwen42/309/head 2025-12-04T09:20:54.4712315Z * [new branch] gh/williamwen42/309/orig -> origin/gh/williamwen42/309/orig 2025-12-04T09:20:54.4712464Z * [new branch] gh/williamwen42/310/base -> origin/gh/williamwen42/310/base 2025-12-04T09:20:54.4712614Z * [new branch] gh/williamwen42/310/head -> origin/gh/williamwen42/310/head 2025-12-04T09:20:54.4712777Z * [new branch] gh/williamwen42/310/orig -> origin/gh/williamwen42/310/orig 2025-12-04T09:20:54.4712930Z * [new branch] gh/williamwen42/311/base -> origin/gh/williamwen42/311/base 2025-12-04T09:20:54.4713082Z * [new branch] gh/williamwen42/311/head -> origin/gh/williamwen42/311/head 2025-12-04T09:20:54.4713245Z * [new branch] gh/williamwen42/311/orig -> origin/gh/williamwen42/311/orig 2025-12-04T09:20:54.4713403Z * [new branch] gh/williamwen42/319/base -> origin/gh/williamwen42/319/base 2025-12-04T09:20:54.4713569Z * [new branch] gh/williamwen42/319/head -> origin/gh/williamwen42/319/head 2025-12-04T09:20:54.4713724Z * [new branch] gh/williamwen42/319/orig -> origin/gh/williamwen42/319/orig 2025-12-04T09:20:54.4713878Z * [new branch] gh/williamwen42/325/base -> origin/gh/williamwen42/325/base 2025-12-04T09:20:54.4714310Z * [new branch] gh/williamwen42/325/head -> origin/gh/williamwen42/325/head 2025-12-04T09:20:54.4714903Z * [new branch] gh/williamwen42/325/orig -> origin/gh/williamwen42/325/orig 2025-12-04T09:20:54.4715306Z * [new branch] gh/williamwen42/326/base -> origin/gh/williamwen42/326/base 2025-12-04T09:20:54.4715498Z * [new branch] gh/williamwen42/326/head -> origin/gh/williamwen42/326/head 2025-12-04T09:20:54.4715660Z * [new branch] gh/williamwen42/326/orig -> origin/gh/williamwen42/326/orig 2025-12-04T09:20:54.4719867Z * [new branch] gh/williamwen42/327/base -> origin/gh/williamwen42/327/base 2025-12-04T09:20:54.4720209Z * [new branch] gh/williamwen42/327/head -> origin/gh/williamwen42/327/head 2025-12-04T09:20:54.4720382Z * [new branch] gh/williamwen42/327/orig -> origin/gh/williamwen42/327/orig 2025-12-04T09:20:54.4720556Z * [new branch] gh/williamwen42/328/base -> origin/gh/williamwen42/328/base 2025-12-04T09:20:54.4720809Z * [new branch] gh/williamwen42/328/head -> origin/gh/williamwen42/328/head 2025-12-04T09:20:54.4726657Z * [new branch] gh/williamwen42/328/orig -> origin/gh/williamwen42/328/orig 2025-12-04T09:20:54.4731038Z * [new branch] gh/williamwen42/329/base -> origin/gh/williamwen42/329/base 2025-12-04T09:20:54.4732965Z * [new branch] gh/williamwen42/329/head -> origin/gh/williamwen42/329/head 2025-12-04T09:20:54.4733174Z * [new branch] gh/williamwen42/329/orig -> origin/gh/williamwen42/329/orig 2025-12-04T09:20:54.4733377Z * [new branch] gh/williamwen42/330/base -> origin/gh/williamwen42/330/base 2025-12-04T09:20:54.4733566Z * [new branch] gh/williamwen42/330/head -> origin/gh/williamwen42/330/head 2025-12-04T09:20:54.4733889Z * [new branch] gh/williamwen42/330/orig -> origin/gh/williamwen42/330/orig 2025-12-04T09:20:54.4734082Z * [new branch] gh/williamwen42/331/base -> origin/gh/williamwen42/331/base 2025-12-04T09:20:54.4734262Z * [new branch] gh/williamwen42/331/head -> origin/gh/williamwen42/331/head 2025-12-04T09:20:54.4734509Z * [new branch] gh/williamwen42/331/orig -> origin/gh/williamwen42/331/orig 2025-12-04T09:20:54.4738651Z * [new branch] gh/williamwen42/332/base -> origin/gh/williamwen42/332/base 2025-12-04T09:20:54.4738878Z * [new branch] gh/williamwen42/332/head -> origin/gh/williamwen42/332/head 2025-12-04T09:20:54.4739034Z * [new branch] gh/williamwen42/332/orig -> origin/gh/williamwen42/332/orig 2025-12-04T09:20:54.4739210Z * [new branch] gh/williamwen42/333/base -> origin/gh/williamwen42/333/base 2025-12-04T09:20:54.4739365Z * [new branch] gh/williamwen42/333/head -> origin/gh/williamwen42/333/head 2025-12-04T09:20:54.4739527Z * [new branch] gh/williamwen42/333/orig -> origin/gh/williamwen42/333/orig 2025-12-04T09:20:54.4743537Z * [new branch] gh/williamwen42/334/base -> origin/gh/williamwen42/334/base 2025-12-04T09:20:54.4747807Z * [new branch] gh/williamwen42/334/head -> origin/gh/williamwen42/334/head 2025-12-04T09:20:54.4748003Z * [new branch] gh/williamwen42/334/orig -> origin/gh/williamwen42/334/orig 2025-12-04T09:20:54.4752026Z * [new branch] gh/williamwen42/335/base -> origin/gh/williamwen42/335/base 2025-12-04T09:20:54.4752562Z * [new branch] gh/williamwen42/335/head -> origin/gh/williamwen42/335/head 2025-12-04T09:20:54.4752792Z * [new branch] gh/williamwen42/335/orig -> origin/gh/williamwen42/335/orig 2025-12-04T09:20:54.4752961Z * [new branch] gh/williamwen42/336/base -> origin/gh/williamwen42/336/base 2025-12-04T09:20:54.4753145Z * [new branch] gh/williamwen42/336/head -> origin/gh/williamwen42/336/head 2025-12-04T09:20:54.4753329Z * [new branch] gh/williamwen42/336/orig -> origin/gh/williamwen42/336/orig 2025-12-04T09:20:54.4753492Z * [new branch] gh/williamwen42/337/base -> origin/gh/williamwen42/337/base 2025-12-04T09:20:54.4753662Z * [new branch] gh/williamwen42/337/head -> origin/gh/williamwen42/337/head 2025-12-04T09:20:54.4753825Z * [new branch] gh/williamwen42/337/orig -> origin/gh/williamwen42/337/orig 2025-12-04T09:20:54.4753991Z * [new branch] gh/williamwen42/338/base -> origin/gh/williamwen42/338/base 2025-12-04T09:20:54.4754388Z * [new branch] gh/williamwen42/338/head -> origin/gh/williamwen42/338/head 2025-12-04T09:20:54.4754547Z * [new branch] gh/williamwen42/338/orig -> origin/gh/williamwen42/338/orig 2025-12-04T09:20:54.4754715Z * [new branch] gh/williamwen42/339/base -> origin/gh/williamwen42/339/base 2025-12-04T09:20:54.4754971Z * [new branch] gh/williamwen42/339/head -> origin/gh/williamwen42/339/head 2025-12-04T09:20:54.4755160Z * [new branch] gh/williamwen42/339/orig -> origin/gh/williamwen42/339/orig 2025-12-04T09:20:54.4755331Z * [new branch] gh/williamwen42/340/base -> origin/gh/williamwen42/340/base 2025-12-04T09:20:54.4755497Z * [new branch] gh/williamwen42/340/head -> origin/gh/williamwen42/340/head 2025-12-04T09:20:54.4755659Z * [new branch] gh/williamwen42/340/orig -> origin/gh/williamwen42/340/orig 2025-12-04T09:20:54.4755822Z * [new branch] gh/williamwen42/341/base -> origin/gh/williamwen42/341/base 2025-12-04T09:20:54.4757112Z * [new branch] gh/williamwen42/341/head -> origin/gh/williamwen42/341/head 2025-12-04T09:20:54.4757279Z * [new branch] gh/williamwen42/341/orig -> origin/gh/williamwen42/341/orig 2025-12-04T09:20:54.4757785Z * [new branch] gh/williamwen42/342/base -> origin/gh/williamwen42/342/base 2025-12-04T09:20:54.4757999Z * [new branch] gh/williamwen42/342/head -> origin/gh/williamwen42/342/head 2025-12-04T09:20:54.4758166Z * [new branch] gh/williamwen42/342/orig -> origin/gh/williamwen42/342/orig 2025-12-04T09:20:54.4758349Z * [new branch] gh/williamwen42/343/base -> origin/gh/williamwen42/343/base 2025-12-04T09:20:54.4758523Z * [new branch] gh/williamwen42/343/head -> origin/gh/williamwen42/343/head 2025-12-04T09:20:54.4758696Z * [new branch] gh/williamwen42/343/orig -> origin/gh/williamwen42/343/orig 2025-12-04T09:20:54.4760102Z * [new branch] gh/williamwen42/344/base -> origin/gh/williamwen42/344/base 2025-12-04T09:20:54.4760279Z * [new branch] gh/williamwen42/344/head -> origin/gh/williamwen42/344/head 2025-12-04T09:20:54.4761601Z * [new branch] gh/williamwen42/344/orig -> origin/gh/williamwen42/344/orig 2025-12-04T09:20:54.4762320Z * [new branch] gh/williamwen42/345/base -> origin/gh/williamwen42/345/base 2025-12-04T09:20:54.4762895Z * [new branch] gh/williamwen42/345/head -> origin/gh/williamwen42/345/head 2025-12-04T09:20:54.4766463Z * [new branch] gh/williamwen42/345/orig -> origin/gh/williamwen42/345/orig 2025-12-04T09:20:54.4767031Z * [new branch] gh/williamwen42/346/base -> origin/gh/williamwen42/346/base 2025-12-04T09:20:54.4767211Z * [new branch] gh/williamwen42/346/head -> origin/gh/williamwen42/346/head 2025-12-04T09:20:54.4767390Z * [new branch] gh/williamwen42/346/orig -> origin/gh/williamwen42/346/orig 2025-12-04T09:20:54.4767595Z * [new branch] gh/williamwen42/347/base -> origin/gh/williamwen42/347/base 2025-12-04T09:20:54.4768021Z * [new branch] gh/williamwen42/347/head -> origin/gh/williamwen42/347/head 2025-12-04T09:20:54.4768864Z * [new branch] gh/williamwen42/347/orig -> origin/gh/williamwen42/347/orig 2025-12-04T09:20:54.4772208Z * [new branch] gh/williamwen42/348/base -> origin/gh/williamwen42/348/base 2025-12-04T09:20:54.4772555Z * [new branch] gh/williamwen42/348/head -> origin/gh/williamwen42/348/head 2025-12-04T09:20:54.4772749Z * [new branch] gh/williamwen42/348/orig -> origin/gh/williamwen42/348/orig 2025-12-04T09:20:54.4772907Z * [new branch] gh/williamwen42/349/base -> origin/gh/williamwen42/349/base 2025-12-04T09:20:54.4773094Z * [new branch] gh/williamwen42/349/head -> origin/gh/williamwen42/349/head 2025-12-04T09:20:54.4774195Z * [new branch] gh/williamwen42/349/orig -> origin/gh/williamwen42/349/orig 2025-12-04T09:20:54.4775331Z * [new branch] gh/williamwen42/350/base -> origin/gh/williamwen42/350/base 2025-12-04T09:20:54.4775527Z * [new branch] gh/williamwen42/350/head -> origin/gh/williamwen42/350/head 2025-12-04T09:20:54.4777658Z * [new branch] gh/williamwen42/350/orig -> origin/gh/williamwen42/350/orig 2025-12-04T09:20:54.4777867Z * [new branch] gh/williamwen42/351/base -> origin/gh/williamwen42/351/base 2025-12-04T09:20:54.4778334Z * [new branch] gh/williamwen42/351/head -> origin/gh/williamwen42/351/head 2025-12-04T09:20:54.4781733Z * [new branch] gh/williamwen42/351/orig -> origin/gh/williamwen42/351/orig 2025-12-04T09:20:54.4781947Z * [new branch] gh/williamwen42/352/base -> origin/gh/williamwen42/352/base 2025-12-04T09:20:54.4782112Z * [new branch] gh/williamwen42/352/head -> origin/gh/williamwen42/352/head 2025-12-04T09:20:54.4782290Z * [new branch] gh/williamwen42/352/orig -> origin/gh/williamwen42/352/orig 2025-12-04T09:20:54.4785923Z * [new branch] gh/williamwen42/353/base -> origin/gh/williamwen42/353/base 2025-12-04T09:20:54.4786121Z * [new branch] gh/williamwen42/353/head -> origin/gh/williamwen42/353/head 2025-12-04T09:20:54.4786305Z * [new branch] gh/williamwen42/353/orig -> origin/gh/williamwen42/353/orig 2025-12-04T09:20:54.4786461Z * [new branch] gh/williamwen42/354/base -> origin/gh/williamwen42/354/base 2025-12-04T09:20:54.4787096Z * [new branch] gh/williamwen42/354/head -> origin/gh/williamwen42/354/head 2025-12-04T09:20:54.4787805Z * [new branch] gh/williamwen42/354/orig -> origin/gh/williamwen42/354/orig 2025-12-04T09:20:54.4789020Z * [new branch] gh/williamwen42/355/base -> origin/gh/williamwen42/355/base 2025-12-04T09:20:54.4789667Z * [new branch] gh/williamwen42/355/head -> origin/gh/williamwen42/355/head 2025-12-04T09:20:54.4790292Z * [new branch] gh/williamwen42/355/orig -> origin/gh/williamwen42/355/orig 2025-12-04T09:20:54.4796750Z * [new branch] gh/williamwen42/356/base -> origin/gh/williamwen42/356/base 2025-12-04T09:20:54.4798643Z * [new branch] gh/williamwen42/356/head -> origin/gh/williamwen42/356/head 2025-12-04T09:20:54.4798837Z * [new branch] gh/williamwen42/356/orig -> origin/gh/williamwen42/356/orig 2025-12-04T09:20:54.4799183Z * [new branch] gh/williamwen42/357/base -> origin/gh/williamwen42/357/base 2025-12-04T09:20:54.4799374Z * [new branch] gh/williamwen42/357/head -> origin/gh/williamwen42/357/head 2025-12-04T09:20:54.4799546Z * [new branch] gh/williamwen42/357/orig -> origin/gh/williamwen42/357/orig 2025-12-04T09:20:54.4799707Z * [new branch] gh/williamwen42/358/base -> origin/gh/williamwen42/358/base 2025-12-04T09:20:54.4799876Z * [new branch] gh/williamwen42/358/head -> origin/gh/williamwen42/358/head 2025-12-04T09:20:54.4800045Z * [new branch] gh/williamwen42/358/orig -> origin/gh/williamwen42/358/orig 2025-12-04T09:20:54.4800201Z * [new branch] gh/xmfan/169/base -> origin/gh/xmfan/169/base 2025-12-04T09:20:54.4800798Z * [new branch] gh/xmfan/169/head -> origin/gh/xmfan/169/head 2025-12-04T09:20:54.4801000Z * [new branch] gh/xmfan/170/base -> origin/gh/xmfan/170/base 2025-12-04T09:20:54.4802172Z * [new branch] gh/xmfan/170/head -> origin/gh/xmfan/170/head 2025-12-04T09:20:54.4802717Z * [new branch] gh/xmfan/274/base -> origin/gh/xmfan/274/base 2025-12-04T09:20:54.4804940Z * [new branch] gh/xmfan/274/head -> origin/gh/xmfan/274/head 2025-12-04T09:20:54.4805128Z * [new branch] gh/xmfan/274/orig -> origin/gh/xmfan/274/orig 2025-12-04T09:20:54.4805510Z * [new branch] gh/xmfan/277/base -> origin/gh/xmfan/277/base 2025-12-04T09:20:54.4807201Z * [new branch] gh/xmfan/277/head -> origin/gh/xmfan/277/head 2025-12-04T09:20:54.4812232Z * [new branch] gh/xmfan/277/orig -> origin/gh/xmfan/277/orig 2025-12-04T09:20:54.4812571Z * [new branch] gh/xmfan/301/base -> origin/gh/xmfan/301/base 2025-12-04T09:20:54.4812722Z * [new branch] gh/xmfan/301/head -> origin/gh/xmfan/301/head 2025-12-04T09:20:54.4812869Z * [new branch] gh/xmfan/301/orig -> origin/gh/xmfan/301/orig 2025-12-04T09:20:54.4813004Z * [new branch] gh/xmfan/304/base -> origin/gh/xmfan/304/base 2025-12-04T09:20:54.4813688Z * [new branch] gh/xmfan/304/head -> origin/gh/xmfan/304/head 2025-12-04T09:20:54.4813845Z * [new branch] gh/xmfan/304/orig -> origin/gh/xmfan/304/orig 2025-12-04T09:20:54.4813993Z * [new branch] gh/xmfan/309/base -> origin/gh/xmfan/309/base 2025-12-04T09:20:54.4814138Z * [new branch] gh/xmfan/309/head -> origin/gh/xmfan/309/head 2025-12-04T09:20:54.4814272Z * [new branch] gh/xmfan/309/orig -> origin/gh/xmfan/309/orig 2025-12-04T09:20:54.4814758Z * [new branch] gh/xmfan/310/base -> origin/gh/xmfan/310/base 2025-12-04T09:20:54.4815202Z * [new branch] gh/xmfan/310/head -> origin/gh/xmfan/310/head 2025-12-04T09:20:54.4820398Z * [new branch] gh/xmfan/310/orig -> origin/gh/xmfan/310/orig 2025-12-04T09:20:54.4820706Z * [new branch] gh/xmfan/311/base -> origin/gh/xmfan/311/base 2025-12-04T09:20:54.4821115Z * [new branch] gh/xmfan/311/head -> origin/gh/xmfan/311/head 2025-12-04T09:20:54.4821280Z * [new branch] gh/xmfan/311/orig -> origin/gh/xmfan/311/orig 2025-12-04T09:20:54.4821823Z * [new branch] gh/xmfan/312/base -> origin/gh/xmfan/312/base 2025-12-04T09:20:54.4822002Z * [new branch] gh/xmfan/312/head -> origin/gh/xmfan/312/head 2025-12-04T09:20:54.4822154Z * [new branch] gh/xmfan/312/orig -> origin/gh/xmfan/312/orig 2025-12-04T09:20:54.4825699Z * [new branch] gh/xmfan/313/base -> origin/gh/xmfan/313/base 2025-12-04T09:20:54.4826043Z * [new branch] gh/xmfan/313/head -> origin/gh/xmfan/313/head 2025-12-04T09:20:54.4826205Z * [new branch] gh/xmfan/313/orig -> origin/gh/xmfan/313/orig 2025-12-04T09:20:54.4826390Z * [new branch] gh/xuanzhang816/27/base -> origin/gh/xuanzhang816/27/base 2025-12-04T09:20:54.4826674Z * [new branch] gh/xuanzhang816/27/head -> origin/gh/xuanzhang816/27/head 2025-12-04T09:20:54.4831048Z * [new branch] gh/xuanzhang816/27/orig -> origin/gh/xuanzhang816/27/orig 2025-12-04T09:20:54.4831393Z * [new branch] gh/xuanzhang816/32/base -> origin/gh/xuanzhang816/32/base 2025-12-04T09:20:54.4831578Z * [new branch] gh/xuanzhang816/32/head -> origin/gh/xuanzhang816/32/head 2025-12-04T09:20:54.4831852Z * [new branch] gh/xuanzhang816/32/orig -> origin/gh/xuanzhang816/32/orig 2025-12-04T09:20:54.4832021Z * [new branch] gh/xuanzhang816/33/base -> origin/gh/xuanzhang816/33/base 2025-12-04T09:20:54.4832267Z * [new branch] gh/xuanzhang816/33/head -> origin/gh/xuanzhang816/33/head 2025-12-04T09:20:54.4832438Z * [new branch] gh/xuanzhang816/33/orig -> origin/gh/xuanzhang816/33/orig 2025-12-04T09:20:54.4832685Z * [new branch] gh/xuanzhang816/34/base -> origin/gh/xuanzhang816/34/base 2025-12-04T09:20:54.4832866Z * [new branch] gh/xuanzhang816/34/head -> origin/gh/xuanzhang816/34/head 2025-12-04T09:20:54.4833731Z * [new branch] gh/xuanzhang816/34/orig -> origin/gh/xuanzhang816/34/orig 2025-12-04T09:20:54.4837790Z * [new branch] gh/xuanzhang816/35/base -> origin/gh/xuanzhang816/35/base 2025-12-04T09:20:54.4838118Z * [new branch] gh/xuanzhang816/35/head -> origin/gh/xuanzhang816/35/head 2025-12-04T09:20:54.4838510Z * [new branch] gh/xuanzhang816/35/orig -> origin/gh/xuanzhang816/35/orig 2025-12-04T09:20:54.4838676Z * [new branch] gh/yanbing-j/11/base -> origin/gh/yanbing-j/11/base 2025-12-04T09:20:54.4838963Z * [new branch] gh/yanbing-j/11/head -> origin/gh/yanbing-j/11/head 2025-12-04T09:20:54.4839666Z * [new branch] gh/yanbing-j/11/orig -> origin/gh/yanbing-j/11/orig 2025-12-04T09:20:54.4840016Z * [new branch] gh/yanbing-j/12/base -> origin/gh/yanbing-j/12/base 2025-12-04T09:20:54.4840190Z * [new branch] gh/yanbing-j/12/head -> origin/gh/yanbing-j/12/head 2025-12-04T09:20:54.4840449Z * [new branch] gh/yanbing-j/12/orig -> origin/gh/yanbing-j/12/orig 2025-12-04T09:20:54.4840612Z * [new branch] gh/yanbing-j/13/base -> origin/gh/yanbing-j/13/base 2025-12-04T09:20:54.4841967Z * [new branch] gh/yanbing-j/13/head -> origin/gh/yanbing-j/13/head 2025-12-04T09:20:54.4843011Z * [new branch] gh/yanbing-j/13/orig -> origin/gh/yanbing-j/13/orig 2025-12-04T09:20:54.4848736Z * [new branch] gh/yanbing-j/14/base -> origin/gh/yanbing-j/14/base 2025-12-04T09:20:54.4853749Z * [new branch] gh/yanbing-j/14/head -> origin/gh/yanbing-j/14/head 2025-12-04T09:20:54.4859227Z * [new branch] gh/yanbing-j/14/orig -> origin/gh/yanbing-j/14/orig 2025-12-04T09:20:54.4862470Z * [new branch] gh/yanbing-j/15/base -> origin/gh/yanbing-j/15/base 2025-12-04T09:20:54.4867550Z * [new branch] gh/yanbing-j/15/head -> origin/gh/yanbing-j/15/head 2025-12-04T09:20:54.4873214Z * [new branch] gh/yanbing-j/15/orig -> origin/gh/yanbing-j/15/orig 2025-12-04T09:20:54.4876231Z * [new branch] gh/yanbing-j/18/base -> origin/gh/yanbing-j/18/base 2025-12-04T09:20:54.4876571Z * [new branch] gh/yanbing-j/18/head -> origin/gh/yanbing-j/18/head 2025-12-04T09:20:54.4876744Z * [new branch] gh/yanbing-j/18/orig -> origin/gh/yanbing-j/18/orig 2025-12-04T09:20:54.4876908Z * [new branch] gh/yanbing-j/19/base -> origin/gh/yanbing-j/19/base 2025-12-04T09:20:54.4877066Z * [new branch] gh/yanbing-j/19/head -> origin/gh/yanbing-j/19/head 2025-12-04T09:20:54.4877221Z * [new branch] gh/yanbing-j/19/orig -> origin/gh/yanbing-j/19/orig 2025-12-04T09:20:54.4877383Z * [new branch] gh/yanbing-j/20/base -> origin/gh/yanbing-j/20/base 2025-12-04T09:20:54.4877543Z * [new branch] gh/yanbing-j/20/head -> origin/gh/yanbing-j/20/head 2025-12-04T09:20:54.4877741Z * [new branch] gh/yanbing-j/20/orig -> origin/gh/yanbing-j/20/orig 2025-12-04T09:20:54.4877887Z * [new branch] gh/yanbing-j/21/base -> origin/gh/yanbing-j/21/base 2025-12-04T09:20:54.4878062Z * [new branch] gh/yanbing-j/21/head -> origin/gh/yanbing-j/21/head 2025-12-04T09:20:54.4878204Z * [new branch] gh/yanbing-j/22/base -> origin/gh/yanbing-j/22/base 2025-12-04T09:20:54.4878344Z * [new branch] gh/yanbing-j/22/head -> origin/gh/yanbing-j/22/head 2025-12-04T09:20:54.4878492Z * [new branch] gh/yanbing-j/22/orig -> origin/gh/yanbing-j/22/orig 2025-12-04T09:20:54.4878633Z * [new branch] gh/yanbing-j/23/base -> origin/gh/yanbing-j/23/base 2025-12-04T09:20:54.4878783Z * [new branch] gh/yanbing-j/23/head -> origin/gh/yanbing-j/23/head 2025-12-04T09:20:54.4879102Z * [new branch] gh/yanbing-j/23/orig -> origin/gh/yanbing-j/23/orig 2025-12-04T09:20:54.4879244Z * [new branch] gh/yanbing-j/24/base -> origin/gh/yanbing-j/24/base 2025-12-04T09:20:54.4879398Z * [new branch] gh/yanbing-j/24/head -> origin/gh/yanbing-j/24/head 2025-12-04T09:20:54.4879663Z * [new branch] gh/yanbing-j/24/orig -> origin/gh/yanbing-j/24/orig 2025-12-04T09:20:54.4879814Z * [new branch] gh/yanbing-j/25/base -> origin/gh/yanbing-j/25/base 2025-12-04T09:20:54.4879957Z * [new branch] gh/yanbing-j/25/head -> origin/gh/yanbing-j/25/head 2025-12-04T09:20:54.4880100Z * [new branch] gh/yanbing-j/25/orig -> origin/gh/yanbing-j/25/orig 2025-12-04T09:20:54.4880249Z * [new branch] gh/yanbing-j/26/base -> origin/gh/yanbing-j/26/base 2025-12-04T09:20:54.4880390Z * [new branch] gh/yanbing-j/26/head -> origin/gh/yanbing-j/26/head 2025-12-04T09:20:54.4880541Z * [new branch] gh/yanbing-j/26/orig -> origin/gh/yanbing-j/26/orig 2025-12-04T09:20:54.4880714Z * [new branch] gh/yang-yu-hang/1/base -> origin/gh/yang-yu-hang/1/base 2025-12-04T09:20:54.4880867Z * [new branch] gh/yang-yu-hang/1/head -> origin/gh/yang-yu-hang/1/head 2025-12-04T09:20:54.4881029Z * [new branch] gh/yang-yu-hang/1/orig -> origin/gh/yang-yu-hang/1/orig 2025-12-04T09:20:54.4881174Z * [new branch] gh/yang-yu-hang/2/base -> origin/gh/yang-yu-hang/2/base 2025-12-04T09:20:54.4881323Z * [new branch] gh/yang-yu-hang/2/head -> origin/gh/yang-yu-hang/2/head 2025-12-04T09:20:54.4881476Z * [new branch] gh/yang-yu-hang/2/orig -> origin/gh/yang-yu-hang/2/orig 2025-12-04T09:20:54.4881633Z * [new branch] gh/yang-yu-hang/3/base -> origin/gh/yang-yu-hang/3/base 2025-12-04T09:20:54.4882164Z * [new branch] gh/yang-yu-hang/3/head -> origin/gh/yang-yu-hang/3/head 2025-12-04T09:20:54.4882322Z * [new branch] gh/yang-yu-hang/3/orig -> origin/gh/yang-yu-hang/3/orig 2025-12-04T09:20:54.4882481Z * [new branch] gh/yangw-dev/12/base -> origin/gh/yangw-dev/12/base 2025-12-04T09:20:54.4882635Z * [new branch] gh/yangw-dev/12/head -> origin/gh/yangw-dev/12/head 2025-12-04T09:20:54.4882776Z * [new branch] gh/yangw-dev/12/orig -> origin/gh/yangw-dev/12/orig 2025-12-04T09:20:54.4882925Z * [new branch] gh/yangw-dev/13/base -> origin/gh/yangw-dev/13/base 2025-12-04T09:20:54.4883064Z * [new branch] gh/yangw-dev/13/head -> origin/gh/yangw-dev/13/head 2025-12-04T09:20:54.4883205Z * [new branch] gh/yangw-dev/13/orig -> origin/gh/yangw-dev/13/orig 2025-12-04T09:20:54.4883355Z * [new branch] gh/yangw-dev/14/base -> origin/gh/yangw-dev/14/base 2025-12-04T09:20:54.4883497Z * [new branch] gh/yangw-dev/14/head -> origin/gh/yangw-dev/14/head 2025-12-04T09:20:54.4883650Z * [new branch] gh/yangw-dev/14/orig -> origin/gh/yangw-dev/14/orig 2025-12-04T09:20:54.4889841Z * [new branch] gh/yangw-dev/15/base -> origin/gh/yangw-dev/15/base 2025-12-04T09:20:54.4890223Z * [new branch] gh/yangw-dev/15/head -> origin/gh/yangw-dev/15/head 2025-12-04T09:20:54.4890486Z * [new branch] gh/yangw-dev/15/orig -> origin/gh/yangw-dev/15/orig 2025-12-04T09:20:54.4890653Z * [new branch] gh/yangw-dev/19/base -> origin/gh/yangw-dev/19/base 2025-12-04T09:20:54.4890811Z * [new branch] gh/yangw-dev/19/head -> origin/gh/yangw-dev/19/head 2025-12-04T09:20:54.4890976Z * [new branch] gh/yangw-dev/19/orig -> origin/gh/yangw-dev/19/orig 2025-12-04T09:20:54.4891157Z * [new branch] gh/yangw-dev/26/base -> origin/gh/yangw-dev/26/base 2025-12-04T09:20:54.4891456Z * [new branch] gh/yangw-dev/26/head -> origin/gh/yangw-dev/26/head 2025-12-04T09:20:54.4892061Z * [new branch] gh/yangw-dev/26/orig -> origin/gh/yangw-dev/26/orig 2025-12-04T09:20:54.4892245Z * [new branch] gh/yangw-dev/27/base -> origin/gh/yangw-dev/27/base 2025-12-04T09:20:54.4892527Z * [new branch] gh/yangw-dev/27/head -> origin/gh/yangw-dev/27/head 2025-12-04T09:20:54.4896993Z * [new branch] gh/yangw-dev/27/orig -> origin/gh/yangw-dev/27/orig 2025-12-04T09:20:54.4897185Z * [new branch] gh/ydwu4/292/base -> origin/gh/ydwu4/292/base 2025-12-04T09:20:54.4897791Z * [new branch] gh/ydwu4/292/head -> origin/gh/ydwu4/292/head 2025-12-04T09:20:54.4897974Z * [new branch] gh/ydwu4/292/orig -> origin/gh/ydwu4/292/orig 2025-12-04T09:20:54.4898123Z * [new branch] gh/ydwu4/294/base -> origin/gh/ydwu4/294/base 2025-12-04T09:20:54.4898285Z * [new branch] gh/ydwu4/294/head -> origin/gh/ydwu4/294/head 2025-12-04T09:20:54.4898429Z * [new branch] gh/ydwu4/294/orig -> origin/gh/ydwu4/294/orig 2025-12-04T09:20:54.4898577Z * [new branch] gh/ydwu4/295/base -> origin/gh/ydwu4/295/base 2025-12-04T09:20:54.4898747Z * [new branch] gh/ydwu4/295/head -> origin/gh/ydwu4/295/head 2025-12-04T09:20:54.4903567Z * [new branch] gh/ydwu4/295/orig -> origin/gh/ydwu4/295/orig 2025-12-04T09:20:54.4903737Z * [new branch] gh/ydwu4/296/base -> origin/gh/ydwu4/296/base 2025-12-04T09:20:54.4903886Z * [new branch] gh/ydwu4/296/head -> origin/gh/ydwu4/296/head 2025-12-04T09:20:54.4904030Z * [new branch] gh/ydwu4/296/orig -> origin/gh/ydwu4/296/orig 2025-12-04T09:20:54.4904196Z * [new branch] gh/ydwu4/306/base -> origin/gh/ydwu4/306/base 2025-12-04T09:20:54.4904337Z * [new branch] gh/ydwu4/306/head -> origin/gh/ydwu4/306/head 2025-12-04T09:20:54.4907894Z * [new branch] gh/ydwu4/306/orig -> origin/gh/ydwu4/306/orig 2025-12-04T09:20:54.4908066Z * [new branch] gh/ydwu4/312/base -> origin/gh/ydwu4/312/base 2025-12-04T09:20:54.4908237Z * [new branch] gh/ydwu4/312/head -> origin/gh/ydwu4/312/head 2025-12-04T09:20:54.4908392Z * [new branch] gh/ydwu4/312/orig -> origin/gh/ydwu4/312/orig 2025-12-04T09:20:54.4908537Z * [new branch] gh/ydwu4/322/base -> origin/gh/ydwu4/322/base 2025-12-04T09:20:54.4915400Z * [new branch] gh/ydwu4/322/head -> origin/gh/ydwu4/322/head 2025-12-04T09:20:54.4917521Z * [new branch] gh/ydwu4/322/orig -> origin/gh/ydwu4/322/orig 2025-12-04T09:20:54.4918139Z * [new branch] gh/ydwu4/327/base -> origin/gh/ydwu4/327/base 2025-12-04T09:20:54.4918314Z * [new branch] gh/ydwu4/327/head -> origin/gh/ydwu4/327/head 2025-12-04T09:20:54.4918455Z * [new branch] gh/ydwu4/327/orig -> origin/gh/ydwu4/327/orig 2025-12-04T09:20:54.4918603Z * [new branch] gh/ydwu4/328/base -> origin/gh/ydwu4/328/base 2025-12-04T09:20:54.4918758Z * [new branch] gh/ydwu4/328/head -> origin/gh/ydwu4/328/head 2025-12-04T09:20:54.4918905Z * [new branch] gh/ydwu4/328/orig -> origin/gh/ydwu4/328/orig 2025-12-04T09:20:54.4919049Z * [new branch] gh/ydwu4/329/base -> origin/gh/ydwu4/329/base 2025-12-04T09:20:54.4919183Z * [new branch] gh/ydwu4/329/head -> origin/gh/ydwu4/329/head 2025-12-04T09:20:54.4919325Z * [new branch] gh/ydwu4/329/orig -> origin/gh/ydwu4/329/orig 2025-12-04T09:20:54.4919608Z * [new branch] gh/ydwu4/330/base -> origin/gh/ydwu4/330/base 2025-12-04T09:20:54.4919752Z * [new branch] gh/ydwu4/330/head -> origin/gh/ydwu4/330/head 2025-12-04T09:20:54.4919888Z * [new branch] gh/ydwu4/330/orig -> origin/gh/ydwu4/330/orig 2025-12-04T09:20:54.4920024Z * [new branch] gh/ydwu4/331/base -> origin/gh/ydwu4/331/base 2025-12-04T09:20:54.4920234Z * [new branch] gh/ydwu4/331/head -> origin/gh/ydwu4/331/head 2025-12-04T09:20:54.4920371Z * [new branch] gh/ydwu4/331/orig -> origin/gh/ydwu4/331/orig 2025-12-04T09:20:54.4920516Z * [new branch] gh/ydwu4/332/base -> origin/gh/ydwu4/332/base 2025-12-04T09:20:54.4920657Z * [new branch] gh/ydwu4/332/head -> origin/gh/ydwu4/332/head 2025-12-04T09:20:54.4920789Z * [new branch] gh/ydwu4/332/orig -> origin/gh/ydwu4/332/orig 2025-12-04T09:20:54.4921518Z * [new branch] gh/ydwu4/333/base -> origin/gh/ydwu4/333/base 2025-12-04T09:20:54.4922182Z * [new branch] gh/ydwu4/333/head -> origin/gh/ydwu4/333/head 2025-12-04T09:20:54.4927374Z * [new branch] gh/ydwu4/333/orig -> origin/gh/ydwu4/333/orig 2025-12-04T09:20:54.4929716Z * [new branch] gh/ydwu4/334/base -> origin/gh/ydwu4/334/base 2025-12-04T09:20:54.4930023Z * [new branch] gh/ydwu4/334/head -> origin/gh/ydwu4/334/head 2025-12-04T09:20:54.4936128Z * [new branch] gh/ydwu4/334/orig -> origin/gh/ydwu4/334/orig 2025-12-04T09:20:54.4941245Z * [new branch] gh/ydwu4/335/base -> origin/gh/ydwu4/335/base 2025-12-04T09:20:54.4946304Z * [new branch] gh/ydwu4/335/head -> origin/gh/ydwu4/335/head 2025-12-04T09:20:54.4951967Z * [new branch] gh/ydwu4/335/orig -> origin/gh/ydwu4/335/orig 2025-12-04T09:20:54.4952581Z * [new branch] gh/ydwu4/337/base -> origin/gh/ydwu4/337/base 2025-12-04T09:20:54.4952743Z * [new branch] gh/ydwu4/337/head -> origin/gh/ydwu4/337/head 2025-12-04T09:20:54.4952885Z * [new branch] gh/ydwu4/337/orig -> origin/gh/ydwu4/337/orig 2025-12-04T09:20:54.4953016Z * [new branch] gh/ydwu4/339/base -> origin/gh/ydwu4/339/base 2025-12-04T09:20:54.4953163Z * [new branch] gh/ydwu4/339/head -> origin/gh/ydwu4/339/head 2025-12-04T09:20:54.4953304Z * [new branch] gh/ydwu4/339/orig -> origin/gh/ydwu4/339/orig 2025-12-04T09:20:54.4953434Z * [new branch] gh/yf225/133/base -> origin/gh/yf225/133/base 2025-12-04T09:20:54.4953567Z * [new branch] gh/yf225/133/head -> origin/gh/yf225/133/head 2025-12-04T09:20:54.4953706Z * [new branch] gh/yf225/93/base -> origin/gh/yf225/93/base 2025-12-04T09:20:54.4953838Z * [new branch] gh/yf225/93/head -> origin/gh/yf225/93/head 2025-12-04T09:20:54.4954006Z * [new branch] gh/yifuwang/152/base -> origin/gh/yifuwang/152/base 2025-12-04T09:20:54.4954148Z * [new branch] gh/yifuwang/152/head -> origin/gh/yifuwang/152/head 2025-12-04T09:20:54.4954293Z * [new branch] gh/yifuwang/152/orig -> origin/gh/yifuwang/152/orig 2025-12-04T09:20:54.4954436Z * [new branch] gh/yifuwang/195/base -> origin/gh/yifuwang/195/base 2025-12-04T09:20:54.4954572Z * [new branch] gh/yifuwang/195/head -> origin/gh/yifuwang/195/head 2025-12-04T09:20:54.4954716Z * [new branch] gh/yifuwang/195/orig -> origin/gh/yifuwang/195/orig 2025-12-04T09:20:54.4954861Z * [new branch] gh/yiming0416/1/base -> origin/gh/yiming0416/1/base 2025-12-04T09:20:54.4955004Z * [new branch] gh/yiming0416/1/head -> origin/gh/yiming0416/1/head 2025-12-04T09:20:54.4955357Z * [new branch] gh/yiming0416/2/base -> origin/gh/yiming0416/2/base 2025-12-04T09:20:54.4955496Z * [new branch] gh/yiming0416/2/head -> origin/gh/yiming0416/2/head 2025-12-04T09:20:54.4957741Z * [new branch] gh/yushangdi/1/base -> origin/gh/yushangdi/1/base 2025-12-04T09:20:54.4958058Z * [new branch] gh/yushangdi/1/head -> origin/gh/yushangdi/1/head 2025-12-04T09:20:54.4958597Z * [new branch] gh/yushangdi/10/base -> origin/gh/yushangdi/10/base 2025-12-04T09:20:54.4958786Z * [new branch] gh/yushangdi/10/head -> origin/gh/yushangdi/10/head 2025-12-04T09:20:54.4958935Z * [new branch] gh/yushangdi/10/orig -> origin/gh/yushangdi/10/orig 2025-12-04T09:20:54.4959086Z * [new branch] gh/yushangdi/11/base -> origin/gh/yushangdi/11/base 2025-12-04T09:20:54.4959229Z * [new branch] gh/yushangdi/11/head -> origin/gh/yushangdi/11/head 2025-12-04T09:20:54.4959414Z * [new branch] gh/yushangdi/11/orig -> origin/gh/yushangdi/11/orig 2025-12-04T09:20:54.4959587Z * [new branch] gh/yushangdi/2/base -> origin/gh/yushangdi/2/base 2025-12-04T09:20:54.4959737Z * [new branch] gh/yushangdi/2/head -> origin/gh/yushangdi/2/head 2025-12-04T09:20:54.4962291Z * [new branch] gh/yushangdi/7/base -> origin/gh/yushangdi/7/base 2025-12-04T09:20:54.4962461Z * [new branch] gh/yushangdi/7/head -> origin/gh/yushangdi/7/head 2025-12-04T09:20:54.4963040Z * [new branch] gh/yushangdi/7/orig -> origin/gh/yushangdi/7/orig 2025-12-04T09:20:54.4963450Z * [new branch] gh/yushangdi/8/base -> origin/gh/yushangdi/8/base 2025-12-04T09:20:54.4971234Z * [new branch] gh/yushangdi/8/head -> origin/gh/yushangdi/8/head 2025-12-04T09:20:54.4976204Z * [new branch] gh/yushangdi/8/orig -> origin/gh/yushangdi/8/orig 2025-12-04T09:20:54.4978666Z * [new branch] gh/yushangdi/9/base -> origin/gh/yushangdi/9/base 2025-12-04T09:20:54.4978859Z * [new branch] gh/yushangdi/9/head -> origin/gh/yushangdi/9/head 2025-12-04T09:20:54.4979021Z * [new branch] gh/yushangdi/9/orig -> origin/gh/yushangdi/9/orig 2025-12-04T09:20:54.4979189Z * [new branch] gh/zklaus/19/base -> origin/gh/zklaus/19/base 2025-12-04T09:20:54.4979340Z * [new branch] gh/zklaus/19/head -> origin/gh/zklaus/19/head 2025-12-04T09:20:54.4979483Z * [new branch] gh/zklaus/19/orig -> origin/gh/zklaus/19/orig 2025-12-04T09:20:54.4979623Z * [new branch] gh/zklaus/20/base -> origin/gh/zklaus/20/base 2025-12-04T09:20:54.4979772Z * [new branch] gh/zklaus/20/head -> origin/gh/zklaus/20/head 2025-12-04T09:20:54.4979912Z * [new branch] gh/zklaus/20/orig -> origin/gh/zklaus/20/orig 2025-12-04T09:20:54.4980078Z * [new branch] gh/zklaus/21/base -> origin/gh/zklaus/21/base 2025-12-04T09:20:54.4980217Z * [new branch] gh/zklaus/21/head -> origin/gh/zklaus/21/head 2025-12-04T09:20:54.4980353Z * [new branch] gh/zklaus/21/orig -> origin/gh/zklaus/21/orig 2025-12-04T09:20:54.4980499Z * [new branch] gh/zklaus/22/base -> origin/gh/zklaus/22/base 2025-12-04T09:20:54.4980639Z * [new branch] gh/zklaus/22/head -> origin/gh/zklaus/22/head 2025-12-04T09:20:54.4980777Z * [new branch] gh/zklaus/22/orig -> origin/gh/zklaus/22/orig 2025-12-04T09:20:54.4983180Z * [new branch] gh/zklaus/23/base -> origin/gh/zklaus/23/base 2025-12-04T09:20:54.4983551Z * [new branch] gh/zklaus/23/head -> origin/gh/zklaus/23/head 2025-12-04T09:20:54.4983714Z * [new branch] gh/zklaus/23/orig -> origin/gh/zklaus/23/orig 2025-12-04T09:20:54.4984025Z * [new branch] gh/zklaus/24/base -> origin/gh/zklaus/24/base 2025-12-04T09:20:54.4984165Z * [new branch] gh/zklaus/24/head -> origin/gh/zklaus/24/head 2025-12-04T09:20:54.4984303Z * [new branch] gh/zklaus/24/orig -> origin/gh/zklaus/24/orig 2025-12-04T09:20:54.4988270Z * [new branch] gh/zou3519/1197/base -> origin/gh/zou3519/1197/base 2025-12-04T09:20:54.4988511Z * [new branch] gh/zou3519/1197/head -> origin/gh/zou3519/1197/head 2025-12-04T09:20:54.4989271Z * [new branch] gh/zou3519/1197/orig -> origin/gh/zou3519/1197/orig 2025-12-04T09:20:54.4989463Z * [new branch] gh/zou3519/1199/base -> origin/gh/zou3519/1199/base 2025-12-04T09:20:54.4989725Z * [new branch] gh/zou3519/1199/head -> origin/gh/zou3519/1199/head 2025-12-04T09:20:54.4989880Z * [new branch] gh/zou3519/1199/orig -> origin/gh/zou3519/1199/orig 2025-12-04T09:20:54.4993881Z * [new branch] gh/zou3519/1200/base -> origin/gh/zou3519/1200/base 2025-12-04T09:20:54.4994075Z * [new branch] gh/zou3519/1200/head -> origin/gh/zou3519/1200/head 2025-12-04T09:20:54.4994617Z * [new branch] gh/zou3519/1200/orig -> origin/gh/zou3519/1200/orig 2025-12-04T09:20:54.4994814Z * [new branch] gh/zou3519/1201/base -> origin/gh/zou3519/1201/base 2025-12-04T09:20:54.4994977Z * [new branch] gh/zou3519/1201/head -> origin/gh/zou3519/1201/head 2025-12-04T09:20:54.4995122Z * [new branch] gh/zou3519/1201/orig -> origin/gh/zou3519/1201/orig 2025-12-04T09:20:54.4999345Z * [new branch] gh/zou3519/1202/base -> origin/gh/zou3519/1202/base 2025-12-04T09:20:54.4999532Z * [new branch] gh/zou3519/1202/head -> origin/gh/zou3519/1202/head 2025-12-04T09:20:54.4999711Z * [new branch] gh/zou3519/1202/orig -> origin/gh/zou3519/1202/orig 2025-12-04T09:20:54.4999884Z * [new branch] gh/zpcore/1/base -> origin/gh/zpcore/1/base 2025-12-04T09:20:54.5000031Z * [new branch] gh/zpcore/1/head -> origin/gh/zpcore/1/head 2025-12-04T09:20:54.5001442Z * [new branch] gh/zpcore/11/base -> origin/gh/zpcore/11/base 2025-12-04T09:20:54.5001594Z * [new branch] gh/zpcore/11/head -> origin/gh/zpcore/11/head 2025-12-04T09:20:54.5002090Z * [new branch] gh/zpcore/11/orig -> origin/gh/zpcore/11/orig 2025-12-04T09:20:54.5006436Z * [new branch] gh/zpcore/12/base -> origin/gh/zpcore/12/base 2025-12-04T09:20:54.5006611Z * [new branch] gh/zpcore/12/head -> origin/gh/zpcore/12/head 2025-12-04T09:20:54.5009734Z * [new branch] gh/zpcore/12/orig -> origin/gh/zpcore/12/orig 2025-12-04T09:20:54.5010086Z * [new branch] gh/zpcore/13/base -> origin/gh/zpcore/13/base 2025-12-04T09:20:54.5010258Z * [new branch] gh/zpcore/13/head -> origin/gh/zpcore/13/head 2025-12-04T09:20:54.5010401Z * [new branch] gh/zpcore/13/orig -> origin/gh/zpcore/13/orig 2025-12-04T09:20:54.5010562Z * [new branch] gh/zpcore/14/base -> origin/gh/zpcore/14/base 2025-12-04T09:20:54.5010698Z * [new branch] gh/zpcore/14/head -> origin/gh/zpcore/14/head 2025-12-04T09:20:54.5010836Z * [new branch] gh/zpcore/14/orig -> origin/gh/zpcore/14/orig 2025-12-04T09:20:54.5016410Z * [new branch] gh/zpcore/15/base -> origin/gh/zpcore/15/base 2025-12-04T09:20:54.5016604Z * [new branch] gh/zpcore/15/head -> origin/gh/zpcore/15/head 2025-12-04T09:20:54.5016745Z * [new branch] gh/zpcore/15/orig -> origin/gh/zpcore/15/orig 2025-12-04T09:20:54.5017077Z * [new branch] gh/zpcore/2/base -> origin/gh/zpcore/2/base 2025-12-04T09:20:54.5017222Z * [new branch] gh/zpcore/2/head -> origin/gh/zpcore/2/head 2025-12-04T09:20:54.5017364Z * [new branch] gh/zpcore/21/base -> origin/gh/zpcore/21/base 2025-12-04T09:20:54.5017581Z * [new branch] gh/zpcore/21/head -> origin/gh/zpcore/21/head 2025-12-04T09:20:54.5017718Z * [new branch] gh/zpcore/21/orig -> origin/gh/zpcore/21/orig 2025-12-04T09:20:54.5018007Z * [new branch] gh/zpcore/22/base -> origin/gh/zpcore/22/base 2025-12-04T09:20:54.5018416Z * [new branch] gh/zpcore/22/head -> origin/gh/zpcore/22/head 2025-12-04T09:20:54.5018554Z * [new branch] gh/zpcore/22/orig -> origin/gh/zpcore/22/orig 2025-12-04T09:20:54.5020193Z * [new branch] gh/zpcore/23/base -> origin/gh/zpcore/23/base 2025-12-04T09:20:54.5020511Z * [new branch] gh/zpcore/23/head -> origin/gh/zpcore/23/head 2025-12-04T09:20:54.5020986Z * [new branch] gh/zpcore/23/orig -> origin/gh/zpcore/23/orig 2025-12-04T09:20:54.5023749Z * [new branch] gh/zpcore/24/base -> origin/gh/zpcore/24/base 2025-12-04T09:20:54.5024100Z * [new branch] gh/zpcore/24/head -> origin/gh/zpcore/24/head 2025-12-04T09:20:54.5024298Z * [new branch] gh/zpcore/24/orig -> origin/gh/zpcore/24/orig 2025-12-04T09:20:54.5024549Z * [new branch] gh/zpcore/25/base -> origin/gh/zpcore/25/base 2025-12-04T09:20:54.5025084Z * [new branch] gh/zpcore/25/head -> origin/gh/zpcore/25/head 2025-12-04T09:20:54.5026272Z * [new branch] gh/zpcore/25/orig -> origin/gh/zpcore/25/orig 2025-12-04T09:20:54.5026799Z * [new branch] gh/zpcore/26/base -> origin/gh/zpcore/26/base 2025-12-04T09:20:54.5029509Z * [new branch] gh/zpcore/26/head -> origin/gh/zpcore/26/head 2025-12-04T09:20:54.5029684Z * [new branch] gh/zpcore/26/orig -> origin/gh/zpcore/26/orig 2025-12-04T09:20:54.5029840Z * [new branch] gh/zpcore/27/base -> origin/gh/zpcore/27/base 2025-12-04T09:20:54.5031901Z * [new branch] gh/zpcore/27/head -> origin/gh/zpcore/27/head 2025-12-04T09:20:54.5032080Z * [new branch] gh/zpcore/27/orig -> origin/gh/zpcore/27/orig 2025-12-04T09:20:54.5036874Z * [new branch] gh/zpcore/28/base -> origin/gh/zpcore/28/base 2025-12-04T09:20:54.5037050Z * [new branch] gh/zpcore/28/head -> origin/gh/zpcore/28/head 2025-12-04T09:20:54.5037198Z * [new branch] gh/zpcore/28/orig -> origin/gh/zpcore/28/orig 2025-12-04T09:20:54.5038764Z * [new branch] gh/zpcore/3/base -> origin/gh/zpcore/3/base 2025-12-04T09:20:54.5038943Z * [new branch] gh/zpcore/3/head -> origin/gh/zpcore/3/head 2025-12-04T09:20:54.5039096Z * [new branch] gh/zpcore/4/base -> origin/gh/zpcore/4/base 2025-12-04T09:20:54.5039245Z * [new branch] gh/zpcore/4/head -> origin/gh/zpcore/4/head 2025-12-04T09:20:54.5039400Z * [new branch] gh/zpcore/5/base -> origin/gh/zpcore/5/base 2025-12-04T09:20:54.5040515Z * [new branch] gh/zpcore/5/head -> origin/gh/zpcore/5/head 2025-12-04T09:20:54.5040874Z * [new branch] gh/zpcore/6/base -> origin/gh/zpcore/6/base 2025-12-04T09:20:54.5041128Z * [new branch] gh/zpcore/6/head -> origin/gh/zpcore/6/head 2025-12-04T09:20:54.5049229Z * [new branch] gh/zpcore/7/base -> origin/gh/zpcore/7/base 2025-12-04T09:20:54.5049473Z * [new branch] gh/zpcore/7/head -> origin/gh/zpcore/7/head 2025-12-04T09:20:54.5049952Z * [new branch] gh/zpcore/8/base -> origin/gh/zpcore/8/base 2025-12-04T09:20:54.5050166Z * [new branch] gh/zpcore/8/head -> origin/gh/zpcore/8/head 2025-12-04T09:20:54.5050311Z * [new branch] google-main -> origin/google-main 2025-12-04T09:20:54.5050716Z * [new branch] guangyey/external_stream -> origin/guangyey/external_stream 2025-12-04T09:20:54.5054402Z * [new branch] guangyey/test_2025 -> origin/guangyey/test_2025 2025-12-04T09:20:54.5059557Z * [new branch] guilhermeleobas/cherry-pick-55d87d9dfd9 -> origin/guilhermeleobas/cherry-pick-55d87d9dfd9 2025-12-04T09:20:54.5063802Z * [new branch] hameerabbasi/complex_tensor_subclass -> origin/hameerabbasi/complex_tensor_subclass 2025-12-04T09:20:54.5067963Z * [new branch] hameerabbasi/fix-ctensor-gradcheck-tests -> origin/hameerabbasi/fix-ctensor-gradcheck-tests 2025-12-04T09:20:54.5072345Z * [new branch] hameerabbasi/gradcheck-allclose -> origin/hameerabbasi/gradcheck-allclose 2025-12-04T09:20:54.5072516Z * [new branch] hc_baseline -> origin/hc_baseline 2025-12-04T09:20:54.5072669Z * [new branch] hhh_rand -> origin/hhh_rand 2025-12-04T09:20:54.5072803Z * [new branch] huba/f1 -> origin/huba/f1 2025-12-04T09:20:54.5073204Z * [new branch] increase-timeout-linux-jammy-cuda12_8-py3_10-gcc11-test -> origin/increase-timeout-linux-jammy-cuda12_8-py3_10-gcc11-test 2025-12-04T09:20:54.5073338Z * [new branch] inlining -> origin/inlining 2025-12-04T09:20:54.5073490Z * [new branch] inlining-ezyang -> origin/inlining-ezyang 2025-12-04T09:20:54.5073669Z * [new branch] install-torchao-0.13.0 -> origin/install-torchao-0.13.0 2025-12-04T09:20:54.5074013Z * [new branch] instrument-trunk-pull-linux-with-job-test-filters -> origin/instrument-trunk-pull-linux-with-job-test-filters 2025-12-04T09:20:54.5074176Z * [new branch] invoke-subgraph -> origin/invoke-subgraph 2025-12-04T09:20:54.5074367Z * [new branch] issue#58739 -> origin/issue#58739 2025-12-04T09:20:54.5074529Z * [new branch] jainapurva-patch-1 -> origin/jainapurva-patch-1 2025-12-04T09:20:54.5074667Z * [new branch] jathu/o3 -> origin/jathu/o3 2025-12-04T09:20:54.5074800Z * [new branch] jathu/sve -> origin/jathu/sve 2025-12-04T09:20:54.5075031Z * [new branch] jcaip/test-cusparselt-version-0.6.2 -> origin/jcaip/test-cusparselt-version-0.6.2 2025-12-04T09:20:54.5075241Z * [new branch] jcaip/update-cusparselt-0.6.2 -> origin/jcaip/update-cusparselt-0.6.2 2025-12-04T09:20:54.5075452Z * [new branch] jiannanWang/memorysnapshot_filter -> origin/jiannanWang/memorysnapshot_filter 2025-12-04T09:20:54.5075670Z * [new branch] jiannanWang/profilerstepwarning -> origin/jiannanWang/profilerstepwarning 2025-12-04T09:20:54.5075845Z * [new branch] jithunnair-amd-patch-1 -> origin/jithunnair-amd-patch-1 2025-12-04T09:20:54.5076018Z * [new branch] jithunnair-amd-patch-10 -> origin/jithunnair-amd-patch-10 2025-12-04T09:20:54.5076194Z * [new branch] jithunnair-amd-patch-2 -> origin/jithunnair-amd-patch-2 2025-12-04T09:20:54.5076356Z * [new branch] jithunnair-amd-patch-3 -> origin/jithunnair-amd-patch-3 2025-12-04T09:20:54.5076525Z * [new branch] jithunnair-amd-patch-4 -> origin/jithunnair-amd-patch-4 2025-12-04T09:20:54.5076685Z * [new branch] jithunnair-amd-patch-5 -> origin/jithunnair-amd-patch-5 2025-12-04T09:20:54.5076848Z * [new branch] jithunnair-amd-patch-6 -> origin/jithunnair-amd-patch-6 2025-12-04T09:20:54.5077021Z * [new branch] jithunnair-amd-patch-7 -> origin/jithunnair-amd-patch-7 2025-12-04T09:20:54.5077340Z * [new branch] jithunnair-amd-patch-8 -> origin/jithunnair-amd-patch-8 2025-12-04T09:20:54.5077518Z * [new branch] jithunnair-amd-patch-9 -> origin/jithunnair-amd-patch-9 2025-12-04T09:20:54.5077679Z * [new branch] justinchu/native-qdq -> origin/justinchu/native-qdq 2025-12-04T09:20:54.5077878Z * [new branch] kainan666/xlf_debug -> origin/kainan666/xlf_debug 2025-12-04T09:20:54.5078024Z * [new branch] kainan_test -> origin/kainan_test 2025-12-04T09:20:54.5078527Z * [new branch] larryliu0820-patch-1 -> origin/larryliu0820-patch-1 2025-12-04T09:20:54.5078737Z * [new branch] leslie/test_group_gemm_epilogues -> origin/leslie/test_group_gemm_epilogues 2025-12-04T09:20:54.5079050Z * [new branch] lessw2020/fix_cutlass_cache_error -> origin/lessw2020/fix_cutlass_cache_error 2025-12-04T09:20:54.5079456Z * [new branch] liaoxuan/shm_all_reduce -> origin/liaoxuan/shm_all_reduce 2025-12-04T09:20:54.5082066Z * [new branch] liaoxuan/test_fa_disable_softmax -> origin/liaoxuan/test_fa_disable_softmax 2025-12-04T09:20:54.5082303Z * [new branch] liaoxuan/test_int8_sdpa -> origin/liaoxuan/test_int8_sdpa 2025-12-04T09:20:54.5082480Z * [new branch] llama4-stable -> origin/llama4-stable 2025-12-04T09:20:54.5089910Z * [new branch] lts/release/1.8 -> origin/lts/release/1.8 2025-12-04T09:20:54.5095620Z * [new branch] lucaskabela/#94773 -> origin/lucaskabela/#94773 2025-12-04T09:20:54.5100511Z * [new branch] lucaskabela/fix_164876 -> origin/lucaskabela/fix_164876 2025-12-04T09:20:54.5102919Z * [new branch] lucaskabela/flop_counter -> origin/lucaskabela/flop_counter 2025-12-04T09:20:54.5103364Z * [new branch] lucaskabela/func_under_decomp -> origin/lucaskabela/func_under_decomp 2025-12-04T09:20:54.5103599Z * [new branch] lucaskabela/functional_in_dynamo -> origin/lucaskabela/functional_in_dynamo 2025-12-04T09:20:54.5103841Z * [new branch] lucaskabela/install_params_as_graph_attr -> origin/lucaskabela/install_params_as_graph_attr 2025-12-04T09:20:54.5104073Z * [new branch] lucaskabela/parameters_as_graph_attr -> origin/lucaskabela/parameters_as_graph_attr 2025-12-04T09:20:54.5104484Z * [new branch] lucaskabela/remove_aot_dispatcher_metadata -> origin/lucaskabela/remove_aot_dispatcher_metadata 2025-12-04T09:20:54.5104666Z * [new branch] lucaskabela/rnn_decomp -> origin/lucaskabela/rnn_decomp 2025-12-04T09:20:54.5104947Z * [new branch] lucaskabela/typing_backends -> origin/lucaskabela/typing_backends 2025-12-04T09:20:54.5105589Z * [new branch] lucaskabela/typing_ctx_manager -> origin/lucaskabela/typing_ctx_manager 2025-12-04T09:20:54.5105828Z * [new branch] lucaskabela/typing_nn_module -> origin/lucaskabela/typing_nn_module 2025-12-04T09:20:54.5106025Z * [new branch] lucaskabela/typing_user_defined -> origin/lucaskabela/typing_user_defined 2025-12-04T09:20:54.5106208Z * [new branch] lucaskabela/typing_variables -> origin/lucaskabela/typing_variables 2025-12-04T09:20:54.5106419Z * [new branch] lucaskabela/typing_variables_dicts -> origin/lucaskabela/typing_variables_dicts 2025-12-04T09:20:54.5106653Z * [new branch] lucaskabela/typing_variables_functions -> origin/lucaskabela/typing_variables_functions 2025-12-04T09:20:54.5106850Z * [new branch] lucaskabela/typing_variables_lists -> origin/lucaskabela/typing_variables_lists 2025-12-04T09:20:54.5107002Z * [new branch] lw/torch_box_by_ref -> origin/lw/torch_box_by_ref 2025-12-04T09:20:54.5107136Z * [new branch] main -> origin/main 2025-12-04T09:20:54.5107435Z * [new branch] malfet-patch-1 -> origin/malfet-patch-1 2025-12-04T09:20:54.5107584Z * [new branch] malfet-patch-2 -> origin/malfet-patch-2 2025-12-04T09:20:54.5107716Z * [new branch] malfet-patch-3 -> origin/malfet-patch-3 2025-12-04T09:20:54.5107897Z * [new branch] malfet-patch-4 -> origin/malfet-patch-4 2025-12-04T09:20:54.5108039Z * [new branch] malfet-patch-5 -> origin/malfet-patch-5 2025-12-04T09:20:54.5108168Z * [new branch] malfet-patch-6 -> origin/malfet-patch-6 2025-12-04T09:20:54.5108302Z * [new branch] malfet-patch-7 -> origin/malfet-patch-7 2025-12-04T09:20:54.5108431Z * [new branch] malfet-patch-8 -> origin/malfet-patch-8 2025-12-04T09:20:54.5108581Z * [new branch] malfet/add-3.14-ci -> origin/malfet/add-3.14-ci 2025-12-04T09:20:54.5108888Z * [new branch] malfet/be-do-not-make-typos-in-build-artifacts -> origin/malfet/be-do-not-make-typos-in-build-artifacts 2025-12-04T09:20:54.5109194Z * [new branch] malfet/be-move-more-settings-to-checkout-pytorch -> origin/malfet/be-move-more-settings-to-checkout-pytorch 2025-12-04T09:20:54.5109434Z * [new branch] malfet/be-remove-misisng-neon-headers -> origin/malfet/be-remove-misisng-neon-headers 2025-12-04T09:20:54.5109792Z * [new branch] malfet/mps-implement-col2im -> origin/malfet/mps-implement-col2im 2025-12-04T09:20:54.5110028Z * [new branch] manuel/aoti_metal_shimify-thread_safe -> origin/manuel/aoti_metal_shimify-thread_safe 2025-12-04T09:20:54.5110967Z * [new branch] manuel/inductor_link_openmp -> origin/manuel/inductor_link_openmp 2025-12-04T09:20:54.5114353Z * [new branch] masnesral/metaconda -> origin/masnesral/metaconda 2025-12-04T09:20:54.5114564Z * [new branch] mem_profiler_flaky_fix -> origin/mem_profiler_flaky_fix 2025-12-04T09:20:54.5114744Z * [new branch] mem_profiler_stack_trace -> origin/mem_profiler_stack_trace 2025-12-04T09:20:54.5114906Z * [new branch] memory_profiler_stack -> origin/memory_profiler_stack 2025-12-04T09:20:54.5115088Z * [new branch] metascroy-patch-1 -> origin/metascroy-patch-1 2025-12-04T09:20:54.5115662Z * [new branch] mingw_posix -> origin/mingw_posix 2025-12-04T09:20:54.5117094Z * [new branch] mlazos/S429861-debug -> origin/mlazos/S429861-debug 2025-12-04T09:20:54.5117370Z * [new branch] mlazos/aa -> origin/mlazos/aa 2025-12-04T09:20:54.5118304Z * [new branch] mlazos/acts -> origin/mlazos/acts 2025-12-04T09:20:54.5119550Z * [new branch] mlazos/arg-renames -> origin/mlazos/arg-renames 2025-12-04T09:20:54.5119803Z * [new branch] mlazos/bad-cudagraphs -> origin/mlazos/bad-cudagraphs 2025-12-04T09:20:54.5120171Z * [new branch] mlazos/baseline-graph-breaks -> origin/mlazos/baseline-graph-breaks 2025-12-04T09:20:54.5120795Z * [new branch] mlazos/beta-tensor -> origin/mlazos/beta-tensor 2025-12-04T09:20:54.5122381Z * [new branch] mlazos/buffers -> origin/mlazos/buffers 2025-12-04T09:20:54.5122554Z * [new branch] mlazos/buffers2 -> origin/mlazos/buffers2 2025-12-04T09:20:54.5125752Z * [new branch] mlazos/buffers3 -> origin/mlazos/buffers3 2025-12-04T09:20:54.5128549Z * [new branch] mlazos/bwd -> origin/mlazos/bwd 2025-12-04T09:20:54.5128896Z * [new branch] mlazos/combo-test -> origin/mlazos/combo-test 2025-12-04T09:20:54.5129103Z * [new branch] mlazos/ctx-cleanup -> origin/mlazos/ctx-cleanup 2025-12-04T09:20:54.5129326Z * [new branch] mlazos/cuda-cmd-log -> origin/mlazos/cuda-cmd-log 2025-12-04T09:20:54.5129674Z * [new branch] mlazos/cudagraph-tests -> origin/mlazos/cudagraph-tests 2025-12-04T09:20:54.5129873Z * [new branch] mlazos/cudagraphs-measurement -> origin/mlazos/cudagraphs-measurement 2025-12-04T09:20:54.5130037Z * [new branch] mlazos/cutlass-test -> origin/mlazos/cutlass-test 2025-12-04T09:20:54.5130510Z * [new branch] mlazos/cutlass-topo-bug -> origin/mlazos/cutlass-topo-bug 2025-12-04T09:20:54.5130692Z * [new branch] mlazos/dataclass-proxy -> origin/mlazos/dataclass-proxy 2025-12-04T09:20:54.5137640Z * [new branch] mlazos/dc-attrs -> origin/mlazos/dc-attrs 2025-12-04T09:20:54.5137828Z * [new branch] mlazos/dc-helion -> origin/mlazos/dc-helion 2025-12-04T09:20:54.5137986Z * [new branch] mlazos/dict-fix -> origin/mlazos/dict-fix 2025-12-04T09:20:54.5138156Z * [new branch] mlazos/disable-tf -> origin/mlazos/disable-tf 2025-12-04T09:20:54.5138312Z * [new branch] mlazos/dupe-fix -> origin/mlazos/dupe-fix 2025-12-04T09:20:54.5138457Z * [new branch] mlazos/dyn-batch -> origin/mlazos/dyn-batch 2025-12-04T09:20:54.5138587Z * [new branch] mlazos/evt -> origin/mlazos/evt 2025-12-04T09:20:54.5138772Z * [new branch] mlazos/extract-examples -> origin/mlazos/extract-examples 2025-12-04T09:20:54.5139091Z * [new branch] mlazos/foreach-op -> origin/mlazos/foreach-op 2025-12-04T09:20:54.5139289Z * [new branch] mlazos/fp8 -> origin/mlazos/fp8 2025-12-04T09:20:54.5139438Z * [new branch] mlazos/fp8-bias -> origin/mlazos/fp8-bias 2025-12-04T09:20:54.5140123Z * [new branch] mlazos/fp8-bias-fusion -> origin/mlazos/fp8-bias-fusion 2025-12-04T09:20:54.5140337Z * [new branch] mlazos/fp8-fixes -> origin/mlazos/fp8-fixes 2025-12-04T09:20:54.5140507Z * [new branch] mlazos/freezing -> origin/mlazos/freezing 2025-12-04T09:20:54.5143818Z * [new branch] mlazos/h-comp -> origin/mlazos/h-comp 2025-12-04T09:20:54.5144148Z * [new branch] mlazos/h-comp2 -> origin/mlazos/h-comp2 2025-12-04T09:20:54.5144350Z * [new branch] mlazos/hash-hop -> origin/mlazos/hash-hop 2025-12-04T09:20:54.5144572Z * [new branch] mlazos/hc -> origin/mlazos/hc 2025-12-04T09:20:54.5144856Z * [new branch] mlazos/hc-cycles -> origin/mlazos/hc-cycles 2025-12-04T09:20:54.5144990Z * [new branch] mlazos/hc-fixes -> origin/mlazos/hc-fixes 2025-12-04T09:20:54.5145242Z * [new branch] mlazos/hc-fixes3 -> origin/mlazos/hc-fixes3 2025-12-04T09:20:54.5150521Z * [new branch] mlazos/hc-fixes4 -> origin/mlazos/hc-fixes4 2025-12-04T09:20:54.5150709Z * [new branch] mlazos/hc-hf -> origin/mlazos/hc-hf 2025-12-04T09:20:54.5150846Z * [new branch] mlazos/hc-mut -> origin/mlazos/hc-mut 2025-12-04T09:20:54.5150985Z * [new branch] mlazos/hc10 -> origin/mlazos/hc10 2025-12-04T09:20:54.5151134Z * [new branch] mlazos/hc11 -> origin/mlazos/hc11 2025-12-04T09:20:54.5151265Z * [new branch] mlazos/hc12 -> origin/mlazos/hc12 2025-12-04T09:20:54.5151386Z * [new branch] mlazos/hc13 -> origin/mlazos/hc13 2025-12-04T09:20:54.5152642Z * [new branch] mlazos/hc14 -> origin/mlazos/hc14 2025-12-04T09:20:54.5152821Z * [new branch] mlazos/hc15 -> origin/mlazos/hc15 2025-12-04T09:20:54.5152953Z * [new branch] mlazos/hc2 -> origin/mlazos/hc2 2025-12-04T09:20:54.5153363Z * [new branch] mlazos/hc4 -> origin/mlazos/hc4 2025-12-04T09:20:54.5153494Z * [new branch] mlazos/hc5 -> origin/mlazos/hc5 2025-12-04T09:20:54.5153615Z * [new branch] mlazos/hc6 -> origin/mlazos/hc6 2025-12-04T09:20:54.5153745Z * [new branch] mlazos/hc7 -> origin/mlazos/hc7 2025-12-04T09:20:54.5153933Z * [new branch] mlazos/hc8 -> origin/mlazos/hc8 2025-12-04T09:20:54.5156419Z * [new branch] mlazos/hc9 -> origin/mlazos/hc9 2025-12-04T09:20:54.5156921Z * [new branch] mlazos/hc_baseline2 -> origin/mlazos/hc_baseline2 2025-12-04T09:20:54.5157127Z * [new branch] mlazos/inductor-streams -> origin/mlazos/inductor-streams 2025-12-04T09:20:54.5157272Z * [new branch] mlazos/main -> origin/mlazos/main 2025-12-04T09:20:54.5157449Z * [new branch] mlazos/mcg2 -> origin/mlazos/mcg2 2025-12-04T09:20:54.5157608Z * [new branch] mlazos/meta-guards -> origin/mlazos/meta-guards 2025-12-04T09:20:54.5158298Z * [new branch] mlazos/mlazos/foreach-map-adam -> origin/mlazos/mlazos/foreach-map-adam 2025-12-04T09:20:54.5158535Z * [new branch] mlazos/mlazos/tf-mode-backup -> origin/mlazos/mlazos/tf-mode-backup 2025-12-04T09:20:54.5159128Z * [new branch] mlazos/mod-fix -> origin/mlazos/mod-fix 2025-12-04T09:20:54.5160000Z * [new branch] mlazos/mode-fix -> origin/mlazos/mode-fix 2025-12-04T09:20:54.5160485Z * [new branch] mlazos/offsets -> origin/mlazos/offsets 2025-12-04T09:20:54.5163423Z * [new branch] mlazos/overguarding -> origin/mlazos/overguarding 2025-12-04T09:20:54.5163723Z * [new branch] mlazos/proxy-ctors -> origin/mlazos/proxy-ctors 2025-12-04T09:20:54.5163925Z * [new branch] mlazos/quant-fix -> origin/mlazos/quant-fix 2025-12-04T09:20:54.5164078Z * [new branch] mlazos/resnet-fix -> origin/mlazos/resnet-fix 2025-12-04T09:20:54.5164287Z * [new branch] mlazos/rm-buf-names -> origin/mlazos/rm-buf-names 2025-12-04T09:20:54.5164854Z * [new branch] mlazos/rm-code -> origin/mlazos/rm-code 2025-12-04T09:20:54.5166687Z * [new branch] mlazos/rm-spam -> origin/mlazos/rm-spam 2025-12-04T09:20:54.5166968Z * [new branch] mlazos/rtp -> origin/mlazos/rtp 2025-12-04T09:20:54.5173720Z * [new branch] mlazos/static-idx-dbg -> origin/mlazos/static-idx-dbg 2025-12-04T09:20:54.5179257Z * [new branch] mlazos/static-inputs-log -> origin/mlazos/static-inputs-log 2025-12-04T09:20:54.5185061Z * [new branch] mlazos/stests -> origin/mlazos/stests 2025-12-04T09:20:54.5188624Z * [new branch] mlazos/stream-ops -> origin/mlazos/stream-ops 2025-12-04T09:20:54.5188973Z * [new branch] mlazos/td-fix2 -> origin/mlazos/td-fix2 2025-12-04T09:20:54.5189212Z * [new branch] mlazos/tensor-hasattr2 -> origin/mlazos/tensor-hasattr2 2025-12-04T09:20:54.5189396Z * [new branch] mlazos/test -> origin/mlazos/test 2025-12-04T09:20:54.5189581Z * [new branch] mlazos/tf-mode -> origin/mlazos/tf-mode 2025-12-04T09:20:54.5189747Z * [new branch] mlazos/tf-mode-backup2 -> origin/mlazos/tf-mode-backup2 2025-12-04T09:20:54.5189914Z * [new branch] mlazos/tf-mode-reland -> origin/mlazos/tf-mode-reland 2025-12-04T09:20:54.5190070Z * [new branch] mlazos/tf-mode-reland2 -> origin/mlazos/tf-mode-reland2 2025-12-04T09:20:54.5190381Z * [new branch] mlazos/tf-mode-reland3 -> origin/mlazos/tf-mode-reland3 2025-12-04T09:20:54.5190736Z * [new branch] mlazos/triton-no-epi -> origin/mlazos/triton-no-epi 2025-12-04T09:20:54.5191395Z * [new branch] mlazos/tune-proto -> origin/mlazos/tune-proto 2025-12-04T09:20:54.5191564Z * [new branch] mlazos/tuple-fixes -> origin/mlazos/tuple-fixes 2025-12-04T09:20:54.5191789Z * [new branch] mlazos/tuple-fixes2 -> origin/mlazos/tuple-fixes2 2025-12-04T09:20:54.5191959Z * [new branch] mlazos/tuple-handling -> origin/mlazos/tuple-handling 2025-12-04T09:20:54.5192127Z * [new branch] mlazos/user-stream-base -> origin/mlazos/user-stream-base 2025-12-04T09:20:54.5192277Z * [new branch] mlazos/user-streams -> origin/mlazos/user-streams 2025-12-04T09:20:54.5192472Z * [new branch] mlazos/user-streams-backup -> origin/mlazos/user-streams-backup 2025-12-04T09:20:54.5192642Z * [new branch] mlazos/user-streams-backup2 -> origin/mlazos/user-streams-backup2 2025-12-04T09:20:54.5192794Z * [new branch] mlazos/vary-beta -> origin/mlazos/vary-beta 2025-12-04T09:20:54.5192932Z * [new branch] mlazos/vary-beta2 -> origin/mlazos/vary-beta2 2025-12-04T09:20:54.5193075Z * [new branch] mlazos/weird-perf1 -> origin/mlazos/weird-perf1 2025-12-04T09:20:54.5193230Z * [new branch] mm_out_dtype_compile -> origin/mm_out_dtype_compile 2025-12-04T09:20:54.5193359Z * [new branch] module-shim -> origin/module-shim 2025-12-04T09:20:54.5193489Z * [new branch] move_config -> origin/move_config 2025-12-04T09:20:54.5193629Z * [new branch] msaroufim/reduce -> origin/msaroufim/reduce 2025-12-04T09:20:54.5193764Z * [new branch] mtia/basic-cmake -> origin/mtia/basic-cmake 2025-12-04T09:20:54.5193964Z * [new branch] mwizak/fix-triton-block-shape -> origin/mwizak/fix-triton-block-shape 2025-12-04T09:20:54.5194098Z * [new branch] my_varlen_backup -> origin/my_varlen_backup 2025-12-04T09:20:54.5194251Z * [new branch] nativert_num_outputs -> origin/nativert_num_outputs 2025-12-04T09:20:54.5198212Z * [new branch] new-codegen -> origin/new-codegen 2025-12-04T09:20:54.5198522Z * [new branch] newtest-base -> origin/newtest-base 2025-12-04T09:20:54.5198715Z * [new branch] ngimel/addmm_dtype -> origin/ngimel/addmm_dtype 2025-12-04T09:20:54.5198856Z * [new branch] ngimel/div_inv -> origin/ngimel/div_inv 2025-12-04T09:20:54.5199162Z * [new branch] ngimel/error_index_list -> origin/ngimel/error_index_list 2025-12-04T09:20:54.5199325Z * [new branch] ngimel/gather_grid -> origin/ngimel/gather_grid 2025-12-04T09:20:54.5200088Z * [new branch] ngimel/gather_grid_release -> origin/ngimel/gather_grid_release 2025-12-04T09:20:54.5200284Z * [new branch] ngimel/gg_new -> origin/ngimel/gg_new 2025-12-04T09:20:54.5200702Z * [new branch] ngimel/hostalloc -> origin/ngimel/hostalloc 2025-12-04T09:20:54.5200873Z * [new branch] ngimel/storage_id -> origin/ngimel/storage_id 2025-12-04T09:20:54.5201011Z * [new branch] nightly -> origin/nightly 2025-12-04T09:20:54.5201242Z * [new branch] nikitaved/addmm_1_rowcol_lt_path_check -> origin/nikitaved/addmm_1_rowcol_lt_path_check 2025-12-04T09:20:54.5201864Z * [new branch] nikitaved/addmm_epilogue_fusions_2d_bias -> origin/nikitaved/addmm_epilogue_fusions_2d_bias 2025-12-04T09:20:54.5202401Z * [new branch] nikitaved/addmm_epilogue_fusions_inductor -> origin/nikitaved/addmm_epilogue_fusions_inductor 2025-12-04T09:20:54.5206961Z * [new branch] nikitaved/addmm_epilogue_fusions_scratch -> origin/nikitaved/addmm_epilogue_fusions_scratch 2025-12-04T09:20:54.5207366Z * [new branch] nikitaved/grad_addmm_epilogue_fusions -> origin/nikitaved/grad_addmm_epilogue_fusions 2025-12-04T09:20:54.5207584Z * [new branch] nikitaved/simpler_can_use_32bit_index -> origin/nikitaved/simpler_can_use_32bit_index 2025-12-04T09:20:54.5207816Z * [new branch] nikitaved/test -> origin/nikitaved/test 2025-12-04T09:20:54.5208071Z * [new branch] nmacchioni-perf-test-async-autotune -> origin/nmacchioni-perf-test-async-autotune 2025-12-04T09:20:54.5213088Z * [new branch] no_distributed_log_spew -> origin/no_distributed_log_spew 2025-12-04T09:20:54.5215328Z * [new branch] nofun-hack -> origin/nofun-hack 2025-12-04T09:20:54.5220415Z * [new branch] norm_bench -> origin/norm_bench 2025-12-04T09:20:54.5226595Z * [new branch] nullplay/fuse_matmul -> origin/nullplay/fuse_matmul 2025-12-04T09:20:54.5231533Z * [new branch] nullplay_fuse_matmul -> origin/nullplay_fuse_matmul 2025-12-04T09:20:54.5233404Z * [new branch] optimizer_test -> origin/optimizer_test 2025-12-04T09:20:54.5233708Z * [new branch] orig/release/1.10 -> origin/orig/release/1.10 2025-12-04T09:20:54.5234025Z * [new branch] orig/release/1.11 -> origin/orig/release/1.11 2025-12-04T09:20:54.5234242Z * [new branch] orig/release/1.12 -> origin/orig/release/1.12 2025-12-04T09:20:54.5234394Z * [new branch] orig/release/1.13 -> origin/orig/release/1.13 2025-12-04T09:20:54.5234626Z * [new branch] orig/release/1.6 -> origin/orig/release/1.6 2025-12-04T09:20:54.5234820Z * [new branch] orig/release/1.7 -> origin/orig/release/1.7 2025-12-04T09:20:54.5235001Z * [new branch] orig/release/1.8 -> origin/orig/release/1.8 2025-12-04T09:20:54.5235153Z * [new branch] orig/release/1.9 -> origin/orig/release/1.9 2025-12-04T09:20:54.5235309Z * [new branch] orig/release/2.0 -> origin/orig/release/2.0 2025-12-04T09:20:54.5235523Z * [new branch] orig/release/2.1 -> origin/orig/release/2.1 2025-12-04T09:20:54.5236200Z * [new branch] orig/release/2.2 -> origin/orig/release/2.2 2025-12-04T09:20:54.5236385Z * [new branch] orig/release/2.3 -> origin/orig/release/2.3 2025-12-04T09:20:54.5236529Z * [new branch] orig/release/2.4 -> origin/orig/release/2.4 2025-12-04T09:20:54.5236658Z * [new branch] orig/release/2.5 -> origin/orig/release/2.5 2025-12-04T09:20:54.5236787Z * [new branch] orig/release/2.6 -> origin/orig/release/2.6 2025-12-04T09:20:54.5236922Z * [new branch] orig/release/2.7 -> origin/orig/release/2.7 2025-12-04T09:20:54.5237055Z * [new branch] orig/release/2.8 -> origin/orig/release/2.8 2025-12-04T09:20:54.5237198Z * [new branch] orig/release/2.9 -> origin/orig/release/2.9 2025-12-04T09:20:54.5237386Z * [new branch] origin/gh/fxdawnn/1/base -> origin/origin/gh/fxdawnn/1/base 2025-12-04T09:20:54.5237561Z * [new branch] origin/gh/fxdawnn/1/orig -> origin/origin/gh/fxdawnn/1/orig 2025-12-04T09:20:54.5237728Z * [new branch] origin/gh/zpcore/14/orig -> origin/origin/gh/zpcore/14/orig 2025-12-04T09:20:54.5237881Z * [new branch] oulgen-patch-1 -> origin/oulgen-patch-1 2025-12-04T09:20:54.5238026Z * [new branch] oulgen-patch-2 -> origin/oulgen-patch-2 2025-12-04T09:20:54.5238167Z * [new branch] oulgen-patch-3 -> origin/oulgen-patch-3 2025-12-04T09:20:54.5238297Z * [new branch] oulgen-patch-4 -> origin/oulgen-patch-4 2025-12-04T09:20:54.5238658Z * [new branch] padded-tensor -> origin/padded-tensor 2025-12-04T09:20:54.5238791Z * [new branch] pca2 -> origin/pca2 2025-12-04T09:20:54.5238938Z * [new branch] per_channel_backup -> origin/per_channel_backup 2025-12-04T09:20:54.5239139Z * [new branch] perf_ops -> origin/perf_ops 2025-12-04T09:20:54.5239269Z * [new branch] perf_ops_2_9 -> origin/perf_ops_2_9 2025-12-04T09:20:54.5239418Z * [new branch] pianpwk-patch-1 -> origin/pianpwk-patch-1 2025-12-04T09:20:54.5239584Z * [new branch] pianpwk/__draft_debug_mode -> origin/pianpwk/__draft_debug_mode 2025-12-04T09:20:54.5239801Z * [new branch] pianpwk/_debug_mode_for_triton_draft -> origin/pianpwk/_debug_mode_for_triton_draft 2025-12-04T09:20:54.5239997Z * [new branch] pianpwk/_debug_nn_module_compile -> origin/pianpwk/_debug_nn_module_compile 2025-12-04T09:20:54.5240160Z * [new branch] pianpwk/_draft_triton_11_3 -> origin/pianpwk/_draft_triton_11_3 2025-12-04T09:20:54.5240340Z * [new branch] pianpwk/_manual_bucket_draft -> origin/pianpwk/_manual_bucket_draft 2025-12-04T09:20:54.5240537Z * [new branch] pianpwk/_profile_w_dispatch_keys -> origin/pianpwk/_profile_w_dispatch_keys 2025-12-04T09:20:54.5240938Z * [new branch] pianpwk/_super_draft_debug_mode -> origin/pianpwk/_super_draft_debug_mode 2025-12-04T09:20:54.5242381Z * [new branch] pianpwk/_unbacked_local_shard_size -> origin/pianpwk/_unbacked_local_shard_size 2025-12-04T09:20:54.5242561Z * [new branch] pianpwk/anomaly_tb -> origin/pianpwk/anomaly_tb 2025-12-04T09:20:54.5243218Z * [new branch] pianpwk/auto_fx_annotate -> origin/pianpwk/auto_fx_annotate 2025-12-04T09:20:54.5244094Z * [new branch] pianpwk/backed_size_oblivious_export -> origin/pianpwk/backed_size_oblivious_export 2025-12-04T09:20:54.5244600Z * [new branch] pianpwk/bert_dynamic_perf -> origin/pianpwk/bert_dynamic_perf 2025-12-04T09:20:54.5245727Z * [new branch] pianpwk/debug_fwd_stack_traces -> origin/pianpwk/debug_fwd_stack_traces 2025-12-04T09:20:54.5246361Z * [new branch] pianpwk/debug_hash_tensor -> origin/pianpwk/debug_hash_tensor 2025-12-04T09:20:54.5247034Z * [new branch] pianpwk/debug_mode_annotate -> origin/pianpwk/debug_mode_annotate 2025-12-04T09:20:54.5247721Z * [new branch] pianpwk/debug_mode_defaults -> origin/pianpwk/debug_mode_defaults 2025-12-04T09:20:54.5249027Z * [new branch] pianpwk/debug_mode_hacks -> origin/pianpwk/debug_mode_hacks 2025-12-04T09:20:54.5249546Z * [new branch] pianpwk/debug_mode_opcall_refactor -> origin/pianpwk/debug_mode_opcall_refactor 2025-12-04T09:20:54.5254253Z * [new branch] pianpwk/debug_mode_show_ids -> origin/pianpwk/debug_mode_show_ids 2025-12-04T09:20:54.5254480Z * [new branch] pianpwk/debug_mode_triton -> origin/pianpwk/debug_mode_triton 2025-12-04T09:20:54.5254680Z * [new branch] pianpwk/debug_show_stack_trace -> origin/pianpwk/debug_show_stack_trace 2025-12-04T09:20:54.5254878Z * [new branch] pianpwk/debug_wait_on_collective -> origin/pianpwk/debug_wait_on_collective 2025-12-04T09:20:54.5255103Z * [new branch] pianpwk/debugmode_compile_tf -> origin/pianpwk/debugmode_compile_tf 2025-12-04T09:20:54.5255347Z * [new branch] pianpwk/dispatch_key_debugging_for_debug -> origin/pianpwk/dispatch_key_debugging_for_debug 2025-12-04T09:20:54.5255535Z * [new branch] pianpwk/draft_debug_mode_tfcompile -> origin/pianpwk/draft_debug_mode_tfcompile 2025-12-04T09:20:54.5255717Z * [new branch] pianpwk/draft_multikernel_nn -> origin/pianpwk/draft_multikernel_nn 2025-12-04T09:20:54.5255921Z * [new branch] pianpwk/draft_multikernel_status_10_5 -> origin/pianpwk/draft_multikernel_status_10_5 2025-12-04T09:20:54.5256294Z * [new branch] pianpwk/dtensor_custom_chunk -> origin/pianpwk/dtensor_custom_chunk 2025-12-04T09:20:54.5257154Z * [new branch] pianpwk/dtensor_unbacked_keypath -> origin/pianpwk/dtensor_unbacked_keypath 2025-12-04T09:20:54.5261874Z * [new branch] pianpwk/event_list_tree -> origin/pianpwk/event_list_tree 2025-12-04T09:20:54.5262099Z * [new branch] pianpwk/false_numel_refs -> origin/pianpwk/false_numel_refs 2025-12-04T09:20:54.5262262Z * [new branch] pianpwk/maybe_guard_rel -> origin/pianpwk/maybe_guard_rel 2025-12-04T09:20:54.5262475Z * [new branch] pianpwk/multikernel_hints_draft -> origin/pianpwk/multikernel_hints_draft 2025-12-04T09:20:54.5262690Z * [new branch] pianpwk/no_size_oblivious_slice_scat -> origin/pianpwk/no_size_oblivious_slice_scat 2025-12-04T09:20:54.5262916Z * [new branch] pianpwk/oblivious_reshape_view_better -> origin/pianpwk/oblivious_reshape_view_better 2025-12-04T09:20:54.5263262Z * [new branch] pianpwk/pre_forward_hook -> origin/pianpwk/pre_forward_hook 2025-12-04T09:20:54.5263489Z * [new branch] pianpwk/skip_python_keys_alternate -> origin/pianpwk/skip_python_keys_alternate 2025-12-04T09:20:54.5263998Z * [new branch] pianpwk/skip_python_keys_in_guards -> origin/pianpwk/skip_python_keys_in_guards 2025-12-04T09:20:54.5264817Z * [new branch] pianpwk/sym_tokens_draft -> origin/pianpwk/sym_tokens_draft 2025-12-04T09:20:54.5265376Z * [new branch] pianpwk/symint_one_hot -> origin/pianpwk/symint_one_hot 2025-12-04T09:20:54.5268019Z * [new branch] pianpwk/test_pointwise_guard_or_false -> origin/pianpwk/test_pointwise_guard_or_false 2025-12-04T09:20:54.5268447Z * [new branch] pianpwk/totally_draft_sym_wrap -> origin/pianpwk/totally_draft_sym_wrap 2025-12-04T09:20:54.5268728Z * [new branch] pianpwk/try_dumb_stuff -> origin/pianpwk/try_dumb_stuff 2025-12-04T09:20:54.5268917Z * [new branch] pianpwk/try_dumb_stuff_2 -> origin/pianpwk/try_dumb_stuff_2 2025-12-04T09:20:54.5269138Z * [new branch] pianpwk/unbacked_dtensor_mm -> origin/pianpwk/unbacked_dtensor_mm 2025-12-04T09:20:54.5269750Z * [new branch] pianpwk/unbacked_tracing_12_2 -> origin/pianpwk/unbacked_tracing_12_2 2025-12-04T09:20:54.5270567Z * [new branch] pianpwk/user_symints -> origin/pianpwk/user_symints 2025-12-04T09:20:54.5271160Z * [new branch] pianpwk/wan21_reshape -> origin/pianpwk/wan21_reshape 2025-12-04T09:20:54.5275306Z * [new branch] piz/fix_partial_backward_1112 -> origin/piz/fix_partial_backward_1112 2025-12-04T09:20:54.5275799Z * [new branch] piz/prop_cache_clean -> origin/piz/prop_cache_clean 2025-12-04T09:20:54.5275977Z * [new branch] pool-separate -> origin/pool-separate 2025-12-04T09:20:54.5276135Z * [new branch] pr-156087 -> origin/pr-156087 2025-12-04T09:20:54.5276261Z * [new branch] pr/131860 -> origin/pr/131860 2025-12-04T09:20:54.5276475Z * [new branch] predispatch_to -> origin/predispatch_to 2025-12-04T09:20:54.5279149Z * [new branch] protect-c17 -> origin/protect-c17 2025-12-04T09:20:54.5279348Z * [new branch] pt-opt-cuda3 -> origin/pt-opt-cuda3 2025-12-04T09:20:54.5279661Z * [new branch] python_compiled_autograd -> origin/python_compiled_autograd 2025-12-04T09:20:54.5281415Z * [new branch] q1l1/fix_device_moved_constant_type_unknown -> origin/q1l1/fix_device_moved_constant_type_unknown 2025-12-04T09:20:54.5282023Z * [new branch] q1l1/fix_wrong_default_type_for_kernel_call_args -> origin/q1l1/fix_wrong_default_type_for_kernel_call_args 2025-12-04T09:20:54.5286721Z * [new branch] qchip/export-D54134695 -> origin/qchip/export-D54134695 2025-12-04T09:20:54.5286905Z * [new branch] quote-pytest_cache -> origin/quote-pytest_cache 2025-12-04T09:20:54.5287106Z * [new branch] reland-accgrad-stream-warn -> origin/reland-accgrad-stream-warn 2025-12-04T09:20:54.5289272Z * [new branch] release/1.10 -> origin/release/1.10 2025-12-04T09:20:54.5289422Z * [new branch] release/1.11 -> origin/release/1.11 2025-12-04T09:20:54.5289558Z * [new branch] release/1.12 -> origin/release/1.12 2025-12-04T09:20:54.5289681Z * [new branch] release/1.13 -> origin/release/1.13 2025-12-04T09:20:54.5289812Z * [new branch] release/1.4 -> origin/release/1.4 2025-12-04T09:20:54.5289956Z * [new branch] release/1.4.1 -> origin/release/1.4.1 2025-12-04T09:20:54.5295054Z * [new branch] release/1.5 -> origin/release/1.5 2025-12-04T09:20:54.5295235Z * [new branch] release/1.6 -> origin/release/1.6 2025-12-04T09:20:54.5295373Z * [new branch] release/1.7 -> origin/release/1.7 2025-12-04T09:20:54.5295502Z * [new branch] release/1.8 -> origin/release/1.8 2025-12-04T09:20:54.5295652Z * [new branch] release/1.9 -> origin/release/1.9 2025-12-04T09:20:54.5296257Z * [new branch] release/2.0 -> origin/release/2.0 2025-12-04T09:20:54.5296398Z * [new branch] release/2.1 -> origin/release/2.1 2025-12-04T09:20:54.5296528Z * [new branch] release/2.2 -> origin/release/2.2 2025-12-04T09:20:54.5296656Z * [new branch] release/2.3 -> origin/release/2.3 2025-12-04T09:20:54.5303448Z * [new branch] release/2.4 -> origin/release/2.4 2025-12-04T09:20:54.5303789Z * [new branch] release/2.5 -> origin/release/2.5 2025-12-04T09:20:54.5303985Z * [new branch] release/2.6 -> origin/release/2.6 2025-12-04T09:20:54.5304143Z * [new branch] release/2.7 -> origin/release/2.7 2025-12-04T09:20:54.5304379Z * [new branch] release/2.8 -> origin/release/2.8 2025-12-04T09:20:54.5304575Z * [new branch] release/2.9 -> origin/release/2.9 2025-12-04T09:20:54.5304839Z * [new branch] release_notes -> origin/release_notes 2025-12-04T09:20:54.5309893Z * [new branch] remove_pyinterpreter -> origin/remove_pyinterpreter 2025-12-04T09:20:54.5316162Z * [new branch] replace-pytorch-labs-20250812-195836 -> origin/replace-pytorch-labs-20250812-195836 2025-12-04T09:20:54.5318527Z * [new branch] replace-pytorch-labs-20250812-200248 -> origin/replace-pytorch-labs-20250812-200248 2025-12-04T09:20:54.5318811Z * [new branch] replace-pytorch-labs-20250812-200324 -> origin/replace-pytorch-labs-20250812-200324 2025-12-04T09:20:54.5319049Z * [new branch] replace-pytorch-labs-20250812-204020 -> origin/replace-pytorch-labs-20250812-204020 2025-12-04T09:20:54.5319313Z * [new branch] revert-131069-gh/krzysztofjordan/1/head -> origin/revert-131069-gh/krzysztofjordan/1/head 2025-12-04T09:20:54.5319542Z * [new branch] revert-131469-gh/andrewor14/51/head -> origin/revert-131469-gh/andrewor14/51/head 2025-12-04T09:20:54.5319740Z * [new branch] revert-152361-gh/fadara01/1/head -> origin/revert-152361-gh/fadara01/1/head 2025-12-04T09:20:54.5319927Z * [new branch] revert-156870-gh/skarjala/3/head -> origin/revert-156870-gh/skarjala/3/head 2025-12-04T09:20:54.5320251Z * [new branch] revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ -> origin/revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ 2025-12-04T09:20:54.5320581Z * [new branch] revert-hoo-invoke-subgraph -> origin/revert-hoo-invoke-subgraph 2025-12-04T09:20:54.5320778Z * [new branch] revert_always_build_distributed -> origin/revert_always_build_distributed 2025-12-04T09:20:54.5320920Z * [new branch] rms_norm_patch -> origin/rms_norm_patch 2025-12-04T09:20:54.5321148Z * [new branch] ruisi/fix_all_to_all_estimation -> origin/ruisi/fix_all_to_all_estimation 2025-12-04T09:20:54.5321319Z * [new branch] ruisi/fix_comm_estimation -> origin/ruisi/fix_comm_estimation 2025-12-04T09:20:54.5321513Z * [new branch] ruisi/fix_dynamic_shape_estimation -> origin/ruisi/fix_dynamic_shape_estimation 2025-12-04T09:20:54.5321902Z * [new branch] ruisi/fix_llama3_autobucketing -> origin/ruisi/fix_llama3_autobucketing 2025-12-04T09:20:54.5322107Z * [new branch] ruisi/fix_manual_bucketing_ep_pass -> origin/ruisi/fix_manual_bucketing_ep_pass 2025-12-04T09:20:54.5322276Z * [new branch] ruisi/manual_bucket_pass -> origin/ruisi/manual_bucket_pass 2025-12-04T09:20:54.5322555Z * [new branch] ryanguo99/cleanup-dynamo-expected-failures -> origin/ryanguo99/cleanup-dynamo-expected-failures 2025-12-04T09:20:54.5322730Z * [new branch] ryanguo99/fix-closure-var -> origin/ryanguo99/fix-closure-var 2025-12-04T09:20:54.5322896Z * [new branch] rzou/faketensor_bench -> origin/rzou/faketensor_bench 2025-12-04T09:20:54.5323023Z * [new branch] rzou/njt -> origin/rzou/njt 2025-12-04T09:20:54.5323149Z * [new branch] rzou/pca -> origin/rzou/pca 2025-12-04T09:20:54.5329130Z * [new branch] rzou/realprop -> origin/rzou/realprop 2025-12-04T09:20:54.5334078Z * [new branch] samplevllm -> origin/samplevllm 2025-12-04T09:20:54.5334462Z * [new branch] sanchitintel/weird_thing_with_test_cpu_select_algorithm -> origin/sanchitintel/weird_thing_with_test_cpu_select_algorithm 2025-12-04T09:20:54.5334809Z * [new branch] sapling-pr-archive-SS-JIA -> origin/sapling-pr-archive-SS-JIA 2025-12-04T09:20:54.5335052Z * [new branch] sapling-pr-archive-tushar00jain -> origin/sapling-pr-archive-tushar00jain 2025-12-04T09:20:54.5335337Z * [new branch] save -> origin/save 2025-12-04T09:20:54.5335856Z * [new branch] scaled_mm -> origin/scaled_mm 2025-12-04T09:20:54.5335984Z * [new branch] scan_attempt -> origin/scan_attempt 2025-12-04T09:20:54.5336210Z * [new branch] sdym/2.5.1 -> origin/sdym/2.5.1 2025-12-04T09:20:54.5336432Z * [new branch] sekyondaMeta-dynamoconfig-fix -> origin/sekyondaMeta-dynamoconfig-fix 2025-12-04T09:20:54.5336586Z * [new branch] shengf/fx-xform-perf -> origin/shengf/fx-xform-perf 2025-12-04T09:20:54.5336760Z * [new branch] shoumikhin-patch-1 -> origin/shoumikhin-patch-1 2025-12-04T09:20:54.5337029Z * [new branch] solve-accuracy-fix -> origin/solve-accuracy-fix 2025-12-04T09:20:54.5337204Z * [new branch] some_rocm_inductor_skips -> origin/some_rocm_inductor_skips 2025-12-04T09:20:54.5337469Z * [new branch] soulitzer/stash-tls-ac -> origin/soulitzer/stash-tls-ac 2025-12-04T09:20:54.5337669Z * [new branch] sparse-mm-bf16-support -> origin/sparse-mm-bf16-support 2025-12-04T09:20:54.5337884Z * [new branch] starterTaskUpdate -> origin/starterTaskUpdate 2025-12-04T09:20:54.5338022Z * [new branch] suo -> origin/suo 2025-12-04T09:20:54.5338980Z * [new branch] sve-poc -> origin/sve-poc 2025-12-04T09:20:54.5339482Z * [new branch] switch-bn -> origin/switch-bn 2025-12-04T09:20:54.5344479Z * [new branch] sy_annotation_in_autograd_hop -> origin/sy_annotation_in_autograd_hop 2025-12-04T09:20:54.5344818Z * [new branch] sy_aot_eager_record -> origin/sy_aot_eager_record 2025-12-04T09:20:54.5344987Z * [new branch] sy_custom_bucketing -> origin/sy_custom_bucketing 2025-12-04T09:20:54.5345335Z * [new branch] sy_debug_mode_test -> origin/sy_debug_mode_test 2025-12-04T09:20:54.5345478Z * [new branch] sy_deserialize -> origin/sy_deserialize 2025-12-04T09:20:54.5345615Z * [new branch] sy_dump_gm_code -> origin/sy_dump_gm_code 2025-12-04T09:20:54.5345869Z * [new branch] sy_exp -> origin/sy_exp 2025-12-04T09:20:54.5346094Z * [new branch] sy_export_annotation -> origin/sy_export_annotation 2025-12-04T09:20:54.5347499Z * [new branch] sy_invoke_subgraph -> origin/sy_invoke_subgraph 2025-12-04T09:20:54.5347756Z * [new branch] sy_kernel_bw_name -> origin/sy_kernel_bw_name 2025-12-04T09:20:54.5348038Z * [new branch] sy_multi_arch -> origin/sy_multi_arch 2025-12-04T09:20:54.5348342Z * [new branch] sy_nn_module_stack -> origin/sy_nn_module_stack 2025-12-04T09:20:54.5349886Z * [new branch] sy_original_dtensor -> origin/sy_original_dtensor 2025-12-04T09:20:54.5350231Z * [new branch] sy_profiler_cia -> origin/sy_profiler_cia 2025-12-04T09:20:54.5350475Z * [new branch] symm_mem_sync -> origin/symm_mem_sync 2025-12-04T09:20:54.5353603Z * [new branch] sympy-bottleneck-repro -> origin/sympy-bottleneck-repro 2025-12-04T09:20:54.5353955Z * [new branch] tensordict_integration -> origin/tensordict_integration 2025-12-04T09:20:54.5354141Z * [new branch] test-move-conda-builds -> origin/test-move-conda-builds 2025-12-04T09:20:54.5354446Z * [new branch] test-old -> origin/test-old 2025-12-04T09:20:54.5359804Z * [new branch] test/bmm_heur -> origin/test/bmm_heur 2025-12-04T09:20:54.5360183Z * [new branch] tianren/customOp_autotune_fix -> origin/tianren/customOp_autotune_fix 2025-12-04T09:20:54.5360449Z * [new branch] tianren/customOp_enable_max_autotune -> origin/tianren/customOp_enable_max_autotune 2025-12-04T09:20:54.5360725Z * [new branch] tianren/customOp_fusion -> origin/tianren/customOp_fusion 2025-12-04T09:20:54.5360991Z * [new branch] tianren/customop_collectiveop_benchmark -> origin/tianren/customop_collectiveop_benchmark 2025-12-04T09:20:54.5361901Z * [new branch] tianren/customop_collectiveop_benchmark_fix -> origin/tianren/customop_collectiveop_benchmark_fix 2025-12-04T09:20:54.5362515Z * [new branch] tianren/customop_dynamic_config -> origin/tianren/customop_dynamic_config 2025-12-04T09:20:54.5362753Z * [new branch] tianren/dynamic_range_input -> origin/tianren/dynamic_range_input 2025-12-04T09:20:54.5362966Z * [new branch] tianren/dynamic_range_input_fix -> origin/tianren/dynamic_range_input_fix 2025-12-04T09:20:54.5363191Z * [new branch] tianren/dynamic_range_input_merge -> origin/tianren/dynamic_range_input_merge 2025-12-04T09:20:54.5363644Z * [new branch] tianren/flex_paged_attn_fix_temp -> origin/tianren/flex_paged_attn_fix_temp 2025-12-04T09:20:54.5364999Z * [new branch] tianren/fx_codegen_dump -> origin/tianren/fx_codegen_dump 2025-12-04T09:20:54.5365423Z * [new branch] tianren/symmetric_memory -> origin/tianren/symmetric_memory 2025-12-04T09:20:54.5365888Z * [new branch] tianren/test -> origin/tianren/test 2025-12-04T09:20:54.5367352Z * [new branch] tidy_performance_cyy -> origin/tidy_performance_cyy 2025-12-04T09:20:54.5367682Z * [new branch] tmp -> origin/tmp 2025-12-04T09:20:54.5368220Z * [new branch] torchtitan_ep -> origin/torchtitan_ep 2025-12-04T09:20:54.5369396Z * [new branch] torchtitan_integration -> origin/torchtitan_integration 2025-12-04T09:20:54.5369892Z * [new branch] trace_fsdp_torchtune_lora -> origin/trace_fsdp_torchtune_lora 2025-12-04T09:20:54.5370958Z * [new branch] traceable_fsdp_unit_tests -> origin/traceable_fsdp_unit_tests 2025-12-04T09:20:54.5371487Z * [new branch] tree_loop_vec_base -> origin/tree_loop_vec_base 2025-12-04T09:20:54.5372565Z * [new branch] triton_kernel -> origin/triton_kernel 2025-12-04T09:20:54.5373246Z * [new branch] tt_pkg_1908 -> origin/tt_pkg_1908 2025-12-04T09:20:54.5373967Z * [new branch] type_dec -> origin/type_dec 2025-12-04T09:20:54.5374856Z * [new branch] udate-sphinx-dependancies -> origin/udate-sphinx-dependancies 2025-12-04T09:20:54.5376216Z * [new branch] update-audio-commit-hash/17630256502-1803-1 -> origin/update-audio-commit-hash/17630256502-1803-1 2025-12-04T09:20:54.5376611Z * [new branch] update-audio-commit-hash/19087141161-1916-1 -> origin/update-audio-commit-hash/19087141161-1916-1 2025-12-04T09:20:54.5377737Z * [new branch] update-audio-commit-hash/19250643381-1929-1 -> origin/update-audio-commit-hash/19250643381-1929-1 2025-12-04T09:20:54.5378099Z * [new branch] update-audio-commit-hash/19397724337-1935-1 -> origin/update-audio-commit-hash/19397724337-1935-1 2025-12-04T09:20:54.5378880Z * [new branch] update-audio-commit-hash/19555670148-1941-1 -> origin/update-audio-commit-hash/19555670148-1941-1 2025-12-04T09:20:54.5380308Z * [new branch] update-audio-commit-hash/19750627930-1946-1 -> origin/update-audio-commit-hash/19750627930-1946-1 2025-12-04T09:20:54.5383807Z * [new branch] update-triton-commit-hash/13663274526-1487-2 -> origin/update-triton-commit-hash/13663274526-1487-2 2025-12-04T09:20:54.5384257Z * [new branch] update-vision-commit-hash/19087141161-1916-1 -> origin/update-vision-commit-hash/19087141161-1916-1 2025-12-04T09:20:54.5384572Z * [new branch] update-vision-commit-hash/19184897099-1925-1 -> origin/update-vision-commit-hash/19184897099-1925-1 2025-12-04T09:20:54.5384934Z * [new branch] update-vision-commit-hash/19250643381-1929-1 -> origin/update-vision-commit-hash/19250643381-1929-1 2025-12-04T09:20:54.5385206Z * [new branch] update-vision-commit-hash/19381328640-1934-1 -> origin/update-vision-commit-hash/19381328640-1934-1 2025-12-04T09:20:54.5387608Z * [new branch] update-vision-commit-hash/19485237164-1938-1 -> origin/update-vision-commit-hash/19485237164-1938-1 2025-12-04T09:20:54.5387929Z * [new branch] update-vllm-commit-hash/18451675449-1879-1 -> origin/update-vllm-commit-hash/18451675449-1879-1 2025-12-04T09:20:54.5388331Z * [new branch] update-vllm-dockerfile -> origin/update-vllm-dockerfile 2025-12-04T09:20:54.5394938Z * [new branch] update-xla-commit-hash/19224287370-211-1 -> origin/update-xla-commit-hash/19224287370-211-1 2025-12-04T09:20:54.5395388Z * [new branch] update-xla-commit-hash/19422028566-212-1 -> origin/update-xla-commit-hash/19422028566-212-1 2025-12-04T09:20:54.5395664Z * [new branch] update-xla-commit-hash/19626841311-213-1 -> origin/update-xla-commit-hash/19626841311-213-1 2025-12-04T09:20:54.5395921Z * [new branch] update_docs_torch_multinomial_issue#125388 -> origin/update_docs_torch_multinomial_issue#125388 2025-12-04T09:20:54.5396228Z * [new branch] update_operator_readme -> origin/update_operator_readme 2025-12-04T09:20:54.5396698Z * [new branch] update_slow_tests_1722488736 -> origin/update_slow_tests_1722488736 2025-12-04T09:20:54.5397437Z * [new branch] update_slow_tests_1722879173 -> origin/update_slow_tests_1722879173 2025-12-04T09:20:54.5397794Z * [new branch] update_slow_tests_1762155677 -> origin/update_slow_tests_1762155677 2025-12-04T09:20:54.5398183Z * [new branch] update_slow_tests_1763365283 -> origin/update_slow_tests_1763365283 2025-12-04T09:20:54.5398418Z * [new branch] update_submodule_FBGEMM -> origin/update_submodule_FBGEMM 2025-12-04T09:20:54.5398594Z * [new branch] update_submodule_kineto -> origin/update_submodule_kineto 2025-12-04T09:20:54.5398913Z * [new branch] update_submodule_tensorpipe -> origin/update_submodule_tensorpipe 2025-12-04T09:20:54.5401276Z * [new branch] upload-tests-for-autorevert -> origin/upload-tests-for-autorevert 2025-12-04T09:20:54.5401438Z * [new branch] v0.1.2 -> origin/v0.1.2 2025-12-04T09:20:54.5401596Z * [new branch] v1.0.1 -> origin/v1.0.1 2025-12-04T09:20:54.5401989Z * [new branch] v1.0.3 -> origin/v1.0.3 2025-12-04T09:20:54.5402151Z * [new branch] v1.1.0 -> origin/v1.1.0 2025-12-04T09:20:54.5402294Z * [new branch] v1.2.0 -> origin/v1.2.0 2025-12-04T09:20:54.5402612Z * [new branch] v1.3.0 -> origin/v1.3.0 2025-12-04T09:20:54.5408786Z * [new branch] v1.3.1 -> origin/v1.3.1 2025-12-04T09:20:54.5409182Z * [new branch] validate_fn -> origin/validate_fn 2025-12-04T09:20:54.5409341Z * [new branch] validations_2.6 -> origin/validations_2.6 2025-12-04T09:20:54.5409588Z * [new branch] validations_2.8 -> origin/validations_2.8 2025-12-04T09:20:54.5409962Z * [new branch] varlen-api -> origin/varlen-api 2025-12-04T09:20:54.5410137Z * [new branch] varlen-api-backup -> origin/varlen-api-backup 2025-12-04T09:20:54.5410841Z * [new branch] varlen_batch_invariance -> origin/varlen_batch_invariance 2025-12-04T09:20:54.5411191Z * [new branch] viable/strict -> origin/viable/strict 2025-12-04T09:20:54.5411572Z * [new branch] vishal9-team/dtensor_parallelism_toy -> origin/vishal9-team/dtensor_parallelism_toy 2025-12-04T09:20:54.5411730Z * [new branch] vllmbuildci -> origin/vllmbuildci 2025-12-04T09:20:54.5413677Z * [new branch] vllmpin -> origin/vllmpin 2025-12-04T09:20:54.5414048Z * [new branch] vscode-recommend-pyrefly -> origin/vscode-recommend-pyrefly 2025-12-04T09:20:54.5414247Z * [new branch] wdvr-patch-1 -> origin/wdvr-patch-1 2025-12-04T09:20:54.5415794Z * [new branch] wdvr/iss_145259 -> origin/wdvr/iss_145259 2025-12-04T09:20:54.5416104Z * [new branch] whc/pei -> origin/whc/pei 2025-12-04T09:20:54.5419325Z * [new branch] whc/pp_fix -> origin/whc/pp_fix 2025-12-04T09:20:54.5419858Z * [new branch] whc/sharding -> origin/whc/sharding 2025-12-04T09:20:54.5420044Z * [new branch] whc/sharding2 -> origin/whc/sharding2 2025-12-04T09:20:54.5420176Z * [new branch] whc/uneven -> origin/whc/uneven 2025-12-04T09:20:54.5425791Z * [new branch] whc/uneven-merge -> origin/whc/uneven-merge 2025-12-04T09:20:54.5425975Z * [new branch] win_warnings -> origin/win_warnings 2025-12-04T09:20:54.5426148Z * [new branch] windows_libtorch_free -> origin/windows_libtorch_free 2025-12-04T09:20:54.5426286Z * [new branch] xmfan-war -> origin/xmfan-war 2025-12-04T09:20:54.5426588Z * [new branch] xmfan/ca_0516 -> origin/xmfan/ca_0516 2025-12-04T09:20:54.5426738Z * [new branch] xmfan/ca_1051b93192 -> origin/xmfan/ca_1051b93192 2025-12-04T09:20:54.5427096Z * [new branch] xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 -> origin/xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 2025-12-04T09:20:54.5427255Z * [new branch] xmfan/ca_5a2be192d1 -> origin/xmfan/ca_5a2be192d1 2025-12-04T09:20:54.5427399Z * [new branch] xmfan/ca_9d59b516e9 -> origin/xmfan/ca_9d59b516e9 2025-12-04T09:20:54.5427822Z * [new branch] xmfan/ca_apr8 -> origin/xmfan/ca_apr8 2025-12-04T09:20:54.5427975Z * [new branch] xmfan/ca_base -> origin/xmfan/ca_base 2025-12-04T09:20:54.5428115Z * [new branch] xmfan/ca_dynamic -> origin/xmfan/ca_dynamic 2025-12-04T09:20:54.5435621Z * [new branch] xmfan/ca_fix_dyn -> origin/xmfan/ca_fix_dyn 2025-12-04T09:20:54.5437980Z * [new branch] xmfan/ca_fix_lowering -> origin/xmfan/ca_fix_lowering 2025-12-04T09:20:54.5438185Z * [new branch] xmfan/ca_fix_polyfills -> origin/xmfan/ca_fix_polyfills 2025-12-04T09:20:54.5438355Z * [new branch] xmfan/ca_jan3 -> origin/xmfan/ca_jan3 2025-12-04T09:20:54.5438497Z * [new branch] xmfan/ca_jun18 -> origin/xmfan/ca_jun18 2025-12-04T09:20:54.5438628Z * [new branch] xmfan/ca_jun24 -> origin/xmfan/ca_jun24 2025-12-04T09:20:54.5438779Z * [new branch] xmfan/ca_nested -> origin/xmfan/ca_nested 2025-12-04T09:20:54.5438936Z * [new branch] xmfan/ca_overhead -> origin/xmfan/ca_overhead 2025-12-04T09:20:54.5439117Z * [new branch] xmfan/ca_overhead_0eba7e5451 -> origin/xmfan/ca_overhead_0eba7e5451 2025-12-04T09:20:54.5439264Z * [new branch] xmfan/cacu_jun18 -> origin/xmfan/cacu_jun18 2025-12-04T09:20:54.5439398Z * [new branch] xmfan/cacu_jun19 -> origin/xmfan/cacu_jun19 2025-12-04T09:20:54.5439538Z * [new branch] xmfan/cacu_jun4 -> origin/xmfan/cacu_jun4 2025-12-04T09:20:54.5439709Z * [new branch] xmfan/disable_duck_shape -> origin/xmfan/disable_duck_shape 2025-12-04T09:20:54.5439904Z * [new branch] xmfan/fca_cpp_node_passthrough -> origin/xmfan/fca_cpp_node_passthrough 2025-12-04T09:20:54.5440187Z * [new branch] xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 2025-12-04T09:20:54.5440628Z * [new branch] xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 2025-12-04T09:20:54.5441068Z * [new branch] xmfan/single_step -> origin/xmfan/single_step 2025-12-04T09:20:54.5441269Z * [new branch] xmfan/sth_0829 -> origin/xmfan/sth_0829 2025-12-04T09:20:54.5442271Z * [new branch] xmfan/test -> origin/xmfan/test 2025-12-04T09:20:54.5448935Z * [new branch] yguo/debug-0226-constexpr -> origin/yguo/debug-0226-constexpr 2025-12-04T09:20:54.5451445Z * [new branch] yguo/new_latest_changes -> origin/yguo/new_latest_changes 2025-12-04T09:20:54.5455039Z * [new branch] yguo/patch_constexpr_changes -> origin/yguo/patch_constexpr_changes 2025-12-04T09:20:54.5455331Z * [new branch] yiming/bootcamp -> origin/yiming/bootcamp 2025-12-04T09:20:54.5455828Z * [new branch] yiming/run_with_start_end_rng_hop -> origin/yiming/run_with_start_end_rng_hop 2025-12-04T09:20:54.5455978Z * [new branch] yolo-llama3 -> origin/yolo-llama3 2025-12-04T09:20:54.5456130Z * [new branch] zainr/canary-test -> origin/zainr/canary-test 2025-12-04T09:20:54.5456533Z * [new branch] zainr/cleanup-gh-runners -> origin/zainr/cleanup-gh-runners 2025-12-04T09:20:54.5456791Z * [new branch] zainr/pull-migration-c -> origin/zainr/pull-migration-c 2025-12-04T09:20:54.5461501Z * [new branch] zainr/test2 -> origin/zainr/test2 2025-12-04T09:20:54.5464453Z * [new branch] zasdfgbnm-patch-3 -> origin/zasdfgbnm-patch-3 2025-12-04T09:20:54.5464636Z * [new branch] zb2p -> origin/zb2p 2025-12-04T09:20:54.5464865Z * [new branch] zeros-and-scatter-part2 -> origin/zeros-and-scatter-part2 2025-12-04T09:20:54.5465057Z * [new branch] zhxchen17/ci/vllm_lora_oom -> origin/zhxchen17/ci/vllm_lora_oom 2025-12-04T09:20:54.5465263Z * [new branch] zhxchen17/ci/vllm_multimodal_oom -> origin/zhxchen17/ci/vllm_multimodal_oom 2025-12-04T09:20:54.5465432Z * [new branch] zhxchen17/ci/vllm_pin -> origin/zhxchen17/ci/vllm_pin 2025-12-04T09:20:54.5465805Z * [new branch] zhxchen17/dynamo/unsafe_drop_all_guards -> origin/zhxchen17/dynamo/unsafe_drop_all_guards 2025-12-04T09:20:54.5470919Z * [new branch] zhxchen17/export/call_override -> origin/zhxchen17/export/call_override 2025-12-04T09:20:54.5475354Z * [new branch] zhxchen17/export/codemod1 -> origin/zhxchen17/export/codemod1 2025-12-04T09:20:54.5477767Z * [new branch] zhxchen17/export/ctx_return -> origin/zhxchen17/export/ctx_return 2025-12-04T09:20:54.5478130Z * [new branch] zhxchen17/export/disable_side_effect_warn -> origin/zhxchen17/export/disable_side_effect_warn 2025-12-04T09:20:54.5478344Z * [new branch] zhxchen17/export/pytree_check -> origin/zhxchen17/export/pytree_check 2025-12-04T09:20:54.5478533Z * [new branch] zhxchen17/precompile/aoti -> origin/zhxchen17/precompile/aoti 2025-12-04T09:20:54.5478738Z * [new branch] zhxchen17/precompile/globals -> origin/zhxchen17/precompile/globals 2025-12-04T09:20:54.5478974Z * [new branch] zhxchen17/precompile/inductor_guards -> origin/zhxchen17/precompile/inductor_guards 2025-12-04T09:20:54.5479136Z * [new branch] zhxchen17/scratch/0 -> origin/zhxchen17/scratch/0 2025-12-04T09:20:54.5479365Z * [new branch] zhxchen17/torch_export_api_update -> origin/zhxchen17/torch_export_api_update 2025-12-04T09:20:54.5479526Z * [new branch] zhxhcen17/moodycamel -> origin/zhxhcen17/moodycamel 2025-12-04T09:20:54.5479684Z * [new branch] zxiiro/build-times -> origin/zxiiro/build-times 2025-12-04T09:20:54.5479839Z * [new branch] zxiiro/c7i.2xlarge -> origin/zxiiro/c7i.2xlarge 2025-12-04T09:20:54.5480004Z * [new branch] zxiiro/c7i.2xlarge.h100 -> origin/zxiiro/c7i.2xlarge.h100 2025-12-04T09:20:54.5480135Z * [new branch] zxiiro/main -> origin/zxiiro/main 2025-12-04T09:20:54.5480284Z * [new branch] zxiiro/risc64 -> origin/zxiiro/risc64 2025-12-04T09:20:54.5480469Z * [new branch] zxiiro/test-multicloud-arc -> origin/zxiiro/test-multicloud-arc 2025-12-04T09:20:54.5480803Z * [new tag] bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug -> bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug 2025-12-04T09:20:54.5480931Z * [new tag] ci/binaries/77164 -> ci/binaries/77164 2025-12-04T09:20:54.5481053Z * [new tag] ciflow/b200/115316 -> ciflow/b200/115316 2025-12-04T09:20:54.5481178Z * [new tag] ciflow/b200/160685 -> ciflow/b200/160685 2025-12-04T09:20:54.5481294Z * [new tag] ciflow/b200/161607 -> ciflow/b200/161607 2025-12-04T09:20:54.5481415Z * [new tag] ciflow/b200/161938 -> ciflow/b200/161938 2025-12-04T09:20:54.5481850Z * [new tag] ciflow/b200/167207 -> ciflow/b200/167207 2025-12-04T09:20:54.5481983Z * [new tag] ciflow/b200/167989 -> ciflow/b200/167989 2025-12-04T09:20:54.5482103Z * [new tag] ciflow/b200/168096 -> ciflow/b200/168096 2025-12-04T09:20:54.5482217Z * [new tag] ciflow/b200/168175 -> ciflow/b200/168175 2025-12-04T09:20:54.5482416Z * [new tag] ciflow/b200/168195 -> ciflow/b200/168195 2025-12-04T09:20:54.5482542Z * [new tag] ciflow/b200/169200 -> ciflow/b200/169200 2025-12-04T09:20:54.5482656Z * [new tag] ciflow/b200/169216 -> ciflow/b200/169216 2025-12-04T09:20:54.5482777Z * [new tag] ciflow/b200/169380 -> ciflow/b200/169380 2025-12-04T09:20:54.5482892Z * [new tag] ciflow/b200/169412 -> ciflow/b200/169412 2025-12-04T09:20:54.5483005Z * [new tag] ciflow/b200/169470 -> ciflow/b200/169470 2025-12-04T09:20:54.5483131Z * [new tag] ciflow/b200/169471 -> ciflow/b200/169471 2025-12-04T09:20:54.5483243Z * [new tag] ciflow/b200/169472 -> ciflow/b200/169472 2025-12-04T09:20:54.5483369Z * [new tag] ciflow/b200/169514 -> ciflow/b200/169514 2025-12-04T09:20:54.5483486Z * [new tag] ciflow/b200/169517 -> ciflow/b200/169517 2025-12-04T09:20:54.5483839Z * [new tag] ciflow/binaries/165922 -> ciflow/binaries/165922 2025-12-04T09:20:54.5484345Z * [new tag] ciflow/binaries/169510 -> ciflow/binaries/169510 2025-12-04T09:20:54.5485688Z * [new tag] ciflow/binaries_wheel/157994 -> ciflow/binaries_wheel/157994 2025-12-04T09:20:54.5486049Z * [new tag] ciflow/binaries_wheel/166829 -> ciflow/binaries_wheel/166829 2025-12-04T09:20:54.5486268Z * [new tag] ciflow/binaries_wheel/167972 -> ciflow/binaries_wheel/167972 2025-12-04T09:20:54.5486761Z * [new tag] ciflow/binaries_wheel/167981 -> ciflow/binaries_wheel/167981 2025-12-04T09:20:54.5487240Z * [new tag] ciflow/dynamo/167695 -> ciflow/dynamo/167695 2025-12-04T09:20:54.5487699Z * [new tag] ciflow/dynamo/168096 -> ciflow/dynamo/168096 2025-12-04T09:20:54.5488704Z * [new tag] ciflow/dynamo/169525 -> ciflow/dynamo/169525 2025-12-04T09:20:54.5489033Z * [new tag] ciflow/h100-cutlass-backend/161938 -> ciflow/h100-cutlass-backend/161938 2025-12-04T09:20:54.5489332Z * [new tag] ciflow/h100-cutlass-backend/161940 -> ciflow/h100-cutlass-backend/161940 2025-12-04T09:20:54.5491510Z * [new tag] ciflow/h100-distributed/168923 -> ciflow/h100-distributed/168923 2025-12-04T09:20:54.5491697Z * [new tag] ciflow/h100-symm-mem/167552 -> ciflow/h100-symm-mem/167552 2025-12-04T09:20:54.5491842Z * [new tag] ciflow/h100-symm-mem/168129 -> ciflow/h100-symm-mem/168129 2025-12-04T09:20:54.5492019Z * [new tag] ciflow/h100-symm-mem/168917 -> ciflow/h100-symm-mem/168917 2025-12-04T09:20:54.5492193Z * [new tag] ciflow/h100-symm-mem/169156 -> ciflow/h100-symm-mem/169156 2025-12-04T09:20:54.5492694Z * [new tag] ciflow/h100-symm-mem/169200 -> ciflow/h100-symm-mem/169200 2025-12-04T09:20:54.5493159Z * [new tag] ciflow/h100-symm-mem/169216 -> ciflow/h100-symm-mem/169216 2025-12-04T09:20:54.5493692Z * [new tag] ciflow/h100-symm-mem/169338 -> ciflow/h100-symm-mem/169338 2025-12-04T09:20:54.5494287Z * [new tag] ciflow/h100-symm-mem/169355 -> ciflow/h100-symm-mem/169355 2025-12-04T09:20:54.5494710Z * [new tag] ciflow/h100-symm-mem/169543 -> ciflow/h100-symm-mem/169543 2025-12-04T09:20:54.5495420Z * [new tag] ciflow/h100/115316 -> ciflow/h100/115316 2025-12-04T09:20:54.5495704Z * [new tag] ciflow/h100/160685 -> ciflow/h100/160685 2025-12-04T09:20:54.5496148Z * [new tag] ciflow/h100/160729 -> ciflow/h100/160729 2025-12-04T09:20:54.5496621Z * [new tag] ciflow/h100/161607 -> ciflow/h100/161607 2025-12-04T09:20:54.5497092Z * [new tag] ciflow/h100/161938 -> ciflow/h100/161938 2025-12-04T09:20:54.5497605Z * [new tag] ciflow/h100/167207 -> ciflow/h100/167207 2025-12-04T09:20:54.5498014Z * [new tag] ciflow/h100/167989 -> ciflow/h100/167989 2025-12-04T09:20:54.5499525Z * [new tag] ciflow/h100/168096 -> ciflow/h100/168096 2025-12-04T09:20:54.5499803Z * [new tag] ciflow/h100/168175 -> ciflow/h100/168175 2025-12-04T09:20:54.5499995Z * [new tag] ciflow/h100/168195 -> ciflow/h100/168195 2025-12-04T09:20:54.5500127Z * [new tag] ciflow/h100/168980 -> ciflow/h100/168980 2025-12-04T09:20:54.5500597Z * [new tag] ciflow/h100/169200 -> ciflow/h100/169200 2025-12-04T09:20:54.5503204Z * [new tag] ciflow/h100/169216 -> ciflow/h100/169216 2025-12-04T09:20:54.5503472Z * [new tag] ciflow/h100/169380 -> ciflow/h100/169380 2025-12-04T09:20:54.5503729Z * [new tag] ciflow/h100/169412 -> ciflow/h100/169412 2025-12-04T09:20:54.5503841Z * [new tag] ciflow/h100/169470 -> ciflow/h100/169470 2025-12-04T09:20:54.5503947Z * [new tag] ciflow/h100/169471 -> ciflow/h100/169471 2025-12-04T09:20:54.5504222Z * [new tag] ciflow/h100/169472 -> ciflow/h100/169472 2025-12-04T09:20:54.5504769Z * [new tag] ciflow/h100/169514 -> ciflow/h100/169514 2025-12-04T09:20:54.5504952Z * [new tag] ciflow/inductor-cu126/168096 -> ciflow/inductor-cu126/168096 2025-12-04T09:20:54.5507073Z * [new tag] ciflow/inductor-micro-benchmark-cpu-x86/168096 -> ciflow/inductor-micro-benchmark-cpu-x86/168096 2025-12-04T09:20:54.5507501Z * [new tag] ciflow/inductor-micro-benchmark/166165 -> ciflow/inductor-micro-benchmark/166165 2025-12-04T09:20:54.5507729Z * [new tag] ciflow/inductor-micro-benchmark/168096 -> ciflow/inductor-micro-benchmark/168096 2025-12-04T09:20:54.5507986Z * [new tag] ciflow/inductor-perf-compare/168096 -> ciflow/inductor-perf-compare/168096 2025-12-04T09:20:54.5508299Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/168073 -> ciflow/inductor-perf-test-nightly-rocm-mi300/168073 2025-12-04T09:20:54.5508896Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/168096 -> ciflow/inductor-perf-test-nightly-rocm-mi300/168096 2025-12-04T09:20:54.5509291Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/169024 -> ciflow/inductor-perf-test-nightly-rocm-mi300/169024 2025-12-04T09:20:54.5510140Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi355/169024 -> ciflow/inductor-perf-test-nightly-rocm-mi355/169024 2025-12-04T09:20:54.5510431Z * [new tag] ciflow/inductor-perf-test-nightly/168096 -> ciflow/inductor-perf-test-nightly/168096 2025-12-04T09:20:54.5511101Z * [new tag] ciflow/inductor-periodic/168096 -> ciflow/inductor-periodic/168096 2025-12-04T09:20:54.5511352Z * [new tag] ciflow/inductor-periodic/169024 -> ciflow/inductor-periodic/169024 2025-12-04T09:20:54.5511881Z * [new tag] ciflow/inductor-periodic/169425 -> ciflow/inductor-periodic/169425 2025-12-04T09:20:54.5514421Z * [new tag] ciflow/inductor-rocm-mi200/165545 -> ciflow/inductor-rocm-mi200/165545 2025-12-04T09:20:54.5514796Z * [new tag] ciflow/inductor-rocm-mi200/165997 -> ciflow/inductor-rocm-mi200/165997 2025-12-04T09:20:54.5514999Z * [new tag] ciflow/inductor-rocm-mi200/168096 -> ciflow/inductor-rocm-mi200/168096 2025-12-04T09:20:54.5515472Z * [new tag] ciflow/inductor-rocm-mi200/169063 -> ciflow/inductor-rocm-mi200/169063 2025-12-04T09:20:54.5516187Z * [new tag] ciflow/inductor-rocm-mi200/169425 -> ciflow/inductor-rocm-mi200/169425 2025-12-04T09:20:54.5516539Z * [new tag] ciflow/inductor-rocm-mi300/165545 -> ciflow/inductor-rocm-mi300/165545 2025-12-04T09:20:54.5516727Z * [new tag] ciflow/inductor-rocm-mi300/168096 -> ciflow/inductor-rocm-mi300/168096 2025-12-04T09:20:54.5516894Z * [new tag] ciflow/inductor-rocm-mi300/169063 -> ciflow/inductor-rocm-mi300/169063 2025-12-04T09:20:54.5517065Z * [new tag] ciflow/inductor-rocm-mi300/169425 -> ciflow/inductor-rocm-mi300/169425 2025-12-04T09:20:54.5517680Z * [new tag] ciflow/inductor-rocm/162052 -> ciflow/inductor-rocm/162052 2025-12-04T09:20:54.5517832Z * [new tag] ciflow/inductor-rocm/168971 -> ciflow/inductor-rocm/168971 2025-12-04T09:20:54.5518417Z * [new tag] ciflow/inductor-windows/168096 -> ciflow/inductor-windows/168096 2025-12-04T09:20:54.5519065Z * [new tag] ciflow/inductor/144542 -> ciflow/inductor/144542 2025-12-04T09:20:54.5519431Z * [new tag] ciflow/inductor/146506 -> ciflow/inductor/146506 2025-12-04T09:20:54.5522680Z * [new tag] ciflow/inductor/147990 -> ciflow/inductor/147990 2025-12-04T09:20:54.5523103Z * [new tag] ciflow/inductor/148294 -> ciflow/inductor/148294 2025-12-04T09:20:54.5523272Z * [new tag] ciflow/inductor/148492 -> ciflow/inductor/148492 2025-12-04T09:20:54.5523406Z * [new tag] ciflow/inductor/157149 -> ciflow/inductor/157149 2025-12-04T09:20:54.5523539Z * [new tag] ciflow/inductor/157994 -> ciflow/inductor/157994 2025-12-04T09:20:54.5523677Z * [new tag] ciflow/inductor/160685 -> ciflow/inductor/160685 2025-12-04T09:20:54.5523824Z * [new tag] ciflow/inductor/160686 -> ciflow/inductor/160686 2025-12-04T09:20:54.5523978Z * [new tag] ciflow/inductor/160687 -> ciflow/inductor/160687 2025-12-04T09:20:54.5524103Z * [new tag] ciflow/inductor/160688 -> ciflow/inductor/160688 2025-12-04T09:20:54.5527159Z * [new tag] ciflow/inductor/160706 -> ciflow/inductor/160706 2025-12-04T09:20:54.5527456Z * [new tag] ciflow/inductor/160729 -> ciflow/inductor/160729 2025-12-04T09:20:54.5527603Z * [new tag] ciflow/inductor/161938 -> ciflow/inductor/161938 2025-12-04T09:20:54.5527841Z * [new tag] ciflow/inductor/161939 -> ciflow/inductor/161939 2025-12-04T09:20:54.5527986Z * [new tag] ciflow/inductor/161940 -> ciflow/inductor/161940 2025-12-04T09:20:54.5528152Z * [new tag] ciflow/inductor/162052 -> ciflow/inductor/162052 2025-12-04T09:20:54.5528336Z * [new tag] ciflow/inductor/162275 -> ciflow/inductor/162275 2025-12-04T09:20:54.5528558Z * [new tag] ciflow/inductor/162795 -> ciflow/inductor/162795 2025-12-04T09:20:54.5529250Z * [new tag] ciflow/inductor/163245 -> ciflow/inductor/163245 2025-12-04T09:20:54.5529447Z * [new tag] ciflow/inductor/163335 -> ciflow/inductor/163335 2025-12-04T09:20:54.5535403Z * [new tag] ciflow/inductor/163503 -> ciflow/inductor/163503 2025-12-04T09:20:54.5535579Z * [new tag] ciflow/inductor/163942 -> ciflow/inductor/163942 2025-12-04T09:20:54.5535877Z * [new tag] ciflow/inductor/165270 -> ciflow/inductor/165270 2025-12-04T09:20:54.5536021Z * [new tag] ciflow/inductor/165274 -> ciflow/inductor/165274 2025-12-04T09:20:54.5536150Z * [new tag] ciflow/inductor/165322 -> ciflow/inductor/165322 2025-12-04T09:20:54.5536504Z * [new tag] ciflow/inductor/165597 -> ciflow/inductor/165597 2025-12-04T09:20:54.5536640Z * [new tag] ciflow/inductor/166063 -> ciflow/inductor/166063 2025-12-04T09:20:54.5536765Z * [new tag] ciflow/inductor/166075 -> ciflow/inductor/166075 2025-12-04T09:20:54.5536974Z * [new tag] ciflow/inductor/166165 -> ciflow/inductor/166165 2025-12-04T09:20:54.5537100Z * [new tag] ciflow/inductor/166254 -> ciflow/inductor/166254 2025-12-04T09:20:54.5537224Z * [new tag] ciflow/inductor/166483 -> ciflow/inductor/166483 2025-12-04T09:20:54.5537358Z * [new tag] ciflow/inductor/166494 -> ciflow/inductor/166494 2025-12-04T09:20:54.5537486Z * [new tag] ciflow/inductor/166545 -> ciflow/inductor/166545 2025-12-04T09:20:54.5537614Z * [new tag] ciflow/inductor/166788 -> ciflow/inductor/166788 2025-12-04T09:20:54.5537751Z * [new tag] ciflow/inductor/166846 -> ciflow/inductor/166846 2025-12-04T09:20:54.5537887Z * [new tag] ciflow/inductor/167300 -> ciflow/inductor/167300 2025-12-04T09:20:54.5538017Z * [new tag] ciflow/inductor/167407 -> ciflow/inductor/167407 2025-12-04T09:20:54.5538149Z * [new tag] ciflow/inductor/167536 -> ciflow/inductor/167536 2025-12-04T09:20:54.5538273Z * [new tag] ciflow/inductor/167552 -> ciflow/inductor/167552 2025-12-04T09:20:54.5538404Z * [new tag] ciflow/inductor/167555 -> ciflow/inductor/167555 2025-12-04T09:20:54.5538527Z * [new tag] ciflow/inductor/167583 -> ciflow/inductor/167583 2025-12-04T09:20:54.5538675Z * [new tag] ciflow/inductor/167599 -> ciflow/inductor/167599 2025-12-04T09:20:54.5539236Z * [new tag] ciflow/inductor/167647 -> ciflow/inductor/167647 2025-12-04T09:20:54.5539488Z * [new tag] ciflow/inductor/167677 -> ciflow/inductor/167677 2025-12-04T09:20:54.5539961Z * [new tag] ciflow/inductor/167680 -> ciflow/inductor/167680 2025-12-04T09:20:54.5540498Z * [new tag] ciflow/inductor/167695 -> ciflow/inductor/167695 2025-12-04T09:20:54.5540903Z * [new tag] ciflow/inductor/167742 -> ciflow/inductor/167742 2025-12-04T09:20:54.5541637Z * [new tag] ciflow/inductor/167768 -> ciflow/inductor/167768 2025-12-04T09:20:54.5545026Z * [new tag] ciflow/inductor/167773 -> ciflow/inductor/167773 2025-12-04T09:20:54.5545207Z * [new tag] ciflow/inductor/167781 -> ciflow/inductor/167781 2025-12-04T09:20:54.5545341Z * [new tag] ciflow/inductor/167880 -> ciflow/inductor/167880 2025-12-04T09:20:54.5545479Z * [new tag] ciflow/inductor/167887 -> ciflow/inductor/167887 2025-12-04T09:20:54.5545642Z * [new tag] ciflow/inductor/167972 -> ciflow/inductor/167972 2025-12-04T09:20:54.5545770Z * [new tag] ciflow/inductor/167989 -> ciflow/inductor/167989 2025-12-04T09:20:54.5546043Z * [new tag] ciflow/inductor/168002 -> ciflow/inductor/168002 2025-12-04T09:20:54.5546257Z * [new tag] ciflow/inductor/168050 -> ciflow/inductor/168050 2025-12-04T09:20:54.5546702Z * [new tag] ciflow/inductor/168051 -> ciflow/inductor/168051 2025-12-04T09:20:54.5547131Z * [new tag] ciflow/inductor/168052 -> ciflow/inductor/168052 2025-12-04T09:20:54.5547745Z * [new tag] ciflow/inductor/168073 -> ciflow/inductor/168073 2025-12-04T09:20:54.5548131Z * [new tag] ciflow/inductor/168096 -> ciflow/inductor/168096 2025-12-04T09:20:54.5548608Z * [new tag] ciflow/inductor/168114 -> ciflow/inductor/168114 2025-12-04T09:20:54.5552018Z * [new tag] ciflow/inductor/168115 -> ciflow/inductor/168115 2025-12-04T09:20:54.5552354Z * [new tag] ciflow/inductor/168127 -> ciflow/inductor/168127 2025-12-04T09:20:54.5552502Z * [new tag] ciflow/inductor/168129 -> ciflow/inductor/168129 2025-12-04T09:20:54.5552638Z * [new tag] ciflow/inductor/168157 -> ciflow/inductor/168157 2025-12-04T09:20:54.5552831Z * [new tag] ciflow/inductor/168175 -> ciflow/inductor/168175 2025-12-04T09:20:54.5552965Z * [new tag] ciflow/inductor/168185 -> ciflow/inductor/168185 2025-12-04T09:20:54.5553092Z * [new tag] ciflow/inductor/168195 -> ciflow/inductor/168195 2025-12-04T09:20:54.5553232Z * [new tag] ciflow/inductor/168209 -> ciflow/inductor/168209 2025-12-04T09:20:54.5553363Z * [new tag] ciflow/inductor/168266 -> ciflow/inductor/168266 2025-12-04T09:20:54.5553710Z * [new tag] ciflow/inductor/168316 -> ciflow/inductor/168316 2025-12-04T09:20:54.5554288Z * [new tag] ciflow/inductor/168326 -> ciflow/inductor/168326 2025-12-04T09:20:54.5554737Z * [new tag] ciflow/inductor/168368 -> ciflow/inductor/168368 2025-12-04T09:20:54.5555272Z * [new tag] ciflow/inductor/168894 -> ciflow/inductor/168894 2025-12-04T09:20:54.5555737Z * [new tag] ciflow/inductor/168934 -> ciflow/inductor/168934 2025-12-04T09:20:54.5556197Z * [new tag] ciflow/inductor/168939 -> ciflow/inductor/168939 2025-12-04T09:20:54.5556750Z * [new tag] ciflow/inductor/168946 -> ciflow/inductor/168946 2025-12-04T09:20:54.5557186Z * [new tag] ciflow/inductor/168950 -> ciflow/inductor/168950 2025-12-04T09:20:54.5557715Z * [new tag] ciflow/inductor/168951 -> ciflow/inductor/168951 2025-12-04T09:20:54.5558225Z * [new tag] ciflow/inductor/168952 -> ciflow/inductor/168952 2025-12-04T09:20:54.5558724Z * [new tag] ciflow/inductor/168955 -> ciflow/inductor/168955 2025-12-04T09:20:54.5559300Z * [new tag] ciflow/inductor/168971 -> ciflow/inductor/168971 2025-12-04T09:20:54.5559750Z * [new tag] ciflow/inductor/168979 -> ciflow/inductor/168979 2025-12-04T09:20:54.5560302Z * [new tag] ciflow/inductor/168980 -> ciflow/inductor/168980 2025-12-04T09:20:54.5560888Z * [new tag] ciflow/inductor/168983 -> ciflow/inductor/168983 2025-12-04T09:20:54.5561311Z * [new tag] ciflow/inductor/169006 -> ciflow/inductor/169006 2025-12-04T09:20:54.5562266Z * [new tag] ciflow/inductor/169023 -> ciflow/inductor/169023 2025-12-04T09:20:54.5562401Z * [new tag] ciflow/inductor/169024 -> ciflow/inductor/169024 2025-12-04T09:20:54.5562905Z * [new tag] ciflow/inductor/169025 -> ciflow/inductor/169025 2025-12-04T09:20:54.5564197Z * [new tag] ciflow/inductor/169066 -> ciflow/inductor/169066 2025-12-04T09:20:54.5564344Z * [new tag] ciflow/inductor/169091 -> ciflow/inductor/169091 2025-12-04T09:20:54.5564498Z * [new tag] ciflow/inductor/169102 -> ciflow/inductor/169102 2025-12-04T09:20:54.5567531Z * [new tag] ciflow/inductor/169103 -> ciflow/inductor/169103 2025-12-04T09:20:54.5567727Z * [new tag] ciflow/inductor/169121 -> ciflow/inductor/169121 2025-12-04T09:20:54.5567867Z * [new tag] ciflow/inductor/169134 -> ciflow/inductor/169134 2025-12-04T09:20:54.5568001Z * [new tag] ciflow/inductor/169135 -> ciflow/inductor/169135 2025-12-04T09:20:54.5568148Z * [new tag] ciflow/inductor/169141 -> ciflow/inductor/169141 2025-12-04T09:20:54.5568280Z * [new tag] ciflow/inductor/169151 -> ciflow/inductor/169151 2025-12-04T09:20:54.5568593Z * [new tag] ciflow/inductor/169161 -> ciflow/inductor/169161 2025-12-04T09:20:54.5568731Z * [new tag] ciflow/inductor/169167 -> ciflow/inductor/169167 2025-12-04T09:20:54.5568865Z * [new tag] ciflow/inductor/169177 -> ciflow/inductor/169177 2025-12-04T09:20:54.5569613Z * [new tag] ciflow/inductor/169185 -> ciflow/inductor/169185 2025-12-04T09:20:54.5569812Z * [new tag] ciflow/inductor/169196 -> ciflow/inductor/169196 2025-12-04T09:20:54.5570431Z * [new tag] ciflow/inductor/169200 -> ciflow/inductor/169200 2025-12-04T09:20:54.5570823Z * [new tag] ciflow/inductor/169204 -> ciflow/inductor/169204 2025-12-04T09:20:54.5571412Z * [new tag] ciflow/inductor/169216 -> ciflow/inductor/169216 2025-12-04T09:20:54.5571850Z * [new tag] ciflow/inductor/169219 -> ciflow/inductor/169219 2025-12-04T09:20:54.5572378Z * [new tag] ciflow/inductor/169220 -> ciflow/inductor/169220 2025-12-04T09:20:54.5572903Z * [new tag] ciflow/inductor/169230 -> ciflow/inductor/169230 2025-12-04T09:20:54.5573354Z * [new tag] ciflow/inductor/169242 -> ciflow/inductor/169242 2025-12-04T09:20:54.5573897Z * [new tag] ciflow/inductor/169245 -> ciflow/inductor/169245 2025-12-04T09:20:54.5574754Z * [new tag] ciflow/inductor/169260 -> ciflow/inductor/169260 2025-12-04T09:20:54.5578155Z * [new tag] ciflow/inductor/169282 -> ciflow/inductor/169282 2025-12-04T09:20:54.5578340Z * [new tag] ciflow/inductor/169286 -> ciflow/inductor/169286 2025-12-04T09:20:54.5578486Z * [new tag] ciflow/inductor/169299 -> ciflow/inductor/169299 2025-12-04T09:20:54.5578621Z * [new tag] ciflow/inductor/169304 -> ciflow/inductor/169304 2025-12-04T09:20:54.5578803Z * [new tag] ciflow/inductor/169305 -> ciflow/inductor/169305 2025-12-04T09:20:54.5578936Z * [new tag] ciflow/inductor/169308 -> ciflow/inductor/169308 2025-12-04T09:20:54.5579454Z * [new tag] ciflow/inductor/169319 -> ciflow/inductor/169319 2025-12-04T09:20:54.5579614Z * [new tag] ciflow/inductor/169326 -> ciflow/inductor/169326 2025-12-04T09:20:54.5579772Z * [new tag] ciflow/inductor/169332 -> ciflow/inductor/169332 2025-12-04T09:20:54.5580484Z * [new tag] ciflow/inductor/169333 -> ciflow/inductor/169333 2025-12-04T09:20:54.5581605Z * [new tag] ciflow/inductor/169336 -> ciflow/inductor/169336 2025-12-04T09:20:54.5581872Z * [new tag] ciflow/inductor/169340 -> ciflow/inductor/169340 2025-12-04T09:20:54.5582109Z * [new tag] ciflow/inductor/169341 -> ciflow/inductor/169341 2025-12-04T09:20:54.5582582Z * [new tag] ciflow/inductor/169343 -> ciflow/inductor/169343 2025-12-04T09:20:54.5582755Z * [new tag] ciflow/inductor/169346 -> ciflow/inductor/169346 2025-12-04T09:20:54.5582899Z * [new tag] ciflow/inductor/169348 -> ciflow/inductor/169348 2025-12-04T09:20:54.5584118Z * [new tag] ciflow/inductor/169350 -> ciflow/inductor/169350 2025-12-04T09:20:54.5584301Z * [new tag] ciflow/inductor/169355 -> ciflow/inductor/169355 2025-12-04T09:20:54.5584547Z * [new tag] ciflow/inductor/169370 -> ciflow/inductor/169370 2025-12-04T09:20:54.5585510Z * [new tag] ciflow/inductor/169375 -> ciflow/inductor/169375 2025-12-04T09:20:54.5585748Z * [new tag] ciflow/inductor/169389 -> ciflow/inductor/169389 2025-12-04T09:20:54.5586219Z * [new tag] ciflow/inductor/169391 -> ciflow/inductor/169391 2025-12-04T09:20:54.5586726Z * [new tag] ciflow/inductor/169393 -> ciflow/inductor/169393 2025-12-04T09:20:54.5587259Z * [new tag] ciflow/inductor/169399 -> ciflow/inductor/169399 2025-12-04T09:20:54.5588584Z * [new tag] ciflow/inductor/169400 -> ciflow/inductor/169400 2025-12-04T09:20:54.5589054Z * [new tag] ciflow/inductor/169415 -> ciflow/inductor/169415 2025-12-04T09:20:54.5589339Z * [new tag] ciflow/inductor/169417 -> ciflow/inductor/169417 2025-12-04T09:20:54.5589534Z * [new tag] ciflow/inductor/169418 -> ciflow/inductor/169418 2025-12-04T09:20:54.5590275Z * [new tag] ciflow/inductor/169430 -> ciflow/inductor/169430 2025-12-04T09:20:54.5590727Z * [new tag] ciflow/inductor/169432 -> ciflow/inductor/169432 2025-12-04T09:20:54.5591161Z * [new tag] ciflow/inductor/169436 -> ciflow/inductor/169436 2025-12-04T09:20:54.5593760Z * [new tag] ciflow/inductor/169437 -> ciflow/inductor/169437 2025-12-04T09:20:54.5596432Z * [new tag] ciflow/inductor/169438 -> ciflow/inductor/169438 2025-12-04T09:20:54.5596609Z * [new tag] ciflow/inductor/169441 -> ciflow/inductor/169441 2025-12-04T09:20:54.5596741Z * [new tag] ciflow/inductor/169446 -> ciflow/inductor/169446 2025-12-04T09:20:54.5596884Z * [new tag] ciflow/inductor/169447 -> ciflow/inductor/169447 2025-12-04T09:20:54.5597019Z * [new tag] ciflow/inductor/169452 -> ciflow/inductor/169452 2025-12-04T09:20:54.5597143Z * [new tag] ciflow/inductor/169455 -> ciflow/inductor/169455 2025-12-04T09:20:54.5597277Z * [new tag] ciflow/inductor/169459 -> ciflow/inductor/169459 2025-12-04T09:20:54.5597402Z * [new tag] ciflow/inductor/169463 -> ciflow/inductor/169463 2025-12-04T09:20:54.5597531Z * [new tag] ciflow/inductor/169476 -> ciflow/inductor/169476 2025-12-04T09:20:54.5597668Z * [new tag] ciflow/inductor/169485 -> ciflow/inductor/169485 2025-12-04T09:20:54.5597902Z * [new tag] ciflow/inductor/169493 -> ciflow/inductor/169493 2025-12-04T09:20:54.5598215Z * [new tag] ciflow/inductor/169496 -> ciflow/inductor/169496 2025-12-04T09:20:54.5598814Z * [new tag] ciflow/inductor/169497 -> ciflow/inductor/169497 2025-12-04T09:20:54.5599266Z * [new tag] ciflow/inductor/169503 -> ciflow/inductor/169503 2025-12-04T09:20:54.5599777Z * [new tag] ciflow/inductor/169504 -> ciflow/inductor/169504 2025-12-04T09:20:54.5600632Z * [new tag] ciflow/inductor/169505 -> ciflow/inductor/169505 2025-12-04T09:20:54.5604245Z * [new tag] ciflow/inductor/169508 -> ciflow/inductor/169508 2025-12-04T09:20:54.5604584Z * [new tag] ciflow/inductor/169509 -> ciflow/inductor/169509 2025-12-04T09:20:54.5604777Z * [new tag] ciflow/inductor/169513 -> ciflow/inductor/169513 2025-12-04T09:20:54.5604919Z * [new tag] ciflow/inductor/169514 -> ciflow/inductor/169514 2025-12-04T09:20:54.5605135Z * [new tag] ciflow/inductor/169515 -> ciflow/inductor/169515 2025-12-04T09:20:54.5605285Z * [new tag] ciflow/inductor/169517 -> ciflow/inductor/169517 2025-12-04T09:20:54.5605452Z * [new tag] ciflow/inductor/169519 -> ciflow/inductor/169519 2025-12-04T09:20:54.5605592Z * [new tag] ciflow/inductor/169520 -> ciflow/inductor/169520 2025-12-04T09:20:54.5605828Z * [new tag] ciflow/inductor/169521 -> ciflow/inductor/169521 2025-12-04T09:20:54.5605971Z * [new tag] ciflow/inductor/169524 -> ciflow/inductor/169524 2025-12-04T09:20:54.5606313Z * [new tag] ciflow/inductor/169527 -> ciflow/inductor/169527 2025-12-04T09:20:54.5606785Z * [new tag] ciflow/inductor/169528 -> ciflow/inductor/169528 2025-12-04T09:20:54.5607472Z * [new tag] ciflow/inductor/169532 -> ciflow/inductor/169532 2025-12-04T09:20:54.5607826Z * [new tag] ciflow/inductor/169535 -> ciflow/inductor/169535 2025-12-04T09:20:54.5608308Z * [new tag] ciflow/inductor/169536 -> ciflow/inductor/169536 2025-12-04T09:20:54.5612510Z * [new tag] ciflow/inductor/169547 -> ciflow/inductor/169547 2025-12-04T09:20:54.5612859Z * [new tag] ciflow/inductor/169548 -> ciflow/inductor/169548 2025-12-04T09:20:54.5613045Z * [new tag] ciflow/inductor/169549 -> ciflow/inductor/169549 2025-12-04T09:20:54.5613198Z * [new tag] ciflow/inductor/169551 -> ciflow/inductor/169551 2025-12-04T09:20:54.5613415Z * [new tag] ciflow/inductor/169552 -> ciflow/inductor/169552 2025-12-04T09:20:54.5613576Z * [new tag] ciflow/inductor/169553 -> ciflow/inductor/169553 2025-12-04T09:20:54.5614270Z * [new tag] ciflow/inductor/169557 -> ciflow/inductor/169557 2025-12-04T09:20:54.5614588Z * [new tag] ciflow/inductor/3b9a386 -> ciflow/inductor/3b9a386 2025-12-04T09:20:54.5614739Z * [new tag] ciflow/inductor/3d4b92b -> ciflow/inductor/3d4b92b 2025-12-04T09:20:54.5614895Z * [new tag] ciflow/inductor/d224ac7 -> ciflow/inductor/d224ac7 2025-12-04T09:20:54.5615061Z * [new tag] ciflow/linux-aarch64/157994 -> ciflow/linux-aarch64/157994 2025-12-04T09:20:54.5615346Z * [new tag] ciflow/linux-aarch64/166075 -> ciflow/linux-aarch64/166075 2025-12-04T09:20:54.5615661Z * [new tag] ciflow/linux-aarch64/166876 -> ciflow/linux-aarch64/166876 2025-12-04T09:20:54.5616130Z * [new tag] ciflow/linux-aarch64/167981 -> ciflow/linux-aarch64/167981 2025-12-04T09:20:54.5616743Z * [new tag] ciflow/mps/166254 -> ciflow/mps/166254 2025-12-04T09:20:54.5617150Z * [new tag] ciflow/mps/169017 -> ciflow/mps/169017 2025-12-04T09:20:54.5620858Z * [new tag] ciflow/mps/169372 -> ciflow/mps/169372 2025-12-04T09:20:54.5621220Z * [new tag] ciflow/mps/169478 -> ciflow/mps/169478 2025-12-04T09:20:54.5621399Z * [new tag] ciflow/op-benchmark/157994 -> ciflow/op-benchmark/157994 2025-12-04T09:20:54.5621552Z * [new tag] ciflow/op-benchmark/166075 -> ciflow/op-benchmark/166075 2025-12-04T09:20:54.5621687Z * [new tag] ciflow/op-benchmark/169544 -> ciflow/op-benchmark/169544 2025-12-04T09:20:54.5621883Z * [new tag] ciflow/periodic-rocm-mi200/165997 -> ciflow/periodic-rocm-mi200/165997 2025-12-04T09:20:54.5622061Z * [new tag] ciflow/periodic-rocm-mi200/166517 -> ciflow/periodic-rocm-mi200/166517 2025-12-04T09:20:54.5622799Z * [new tag] ciflow/periodic-rocm-mi200/169063 -> ciflow/periodic-rocm-mi200/169063 2025-12-04T09:20:54.5623009Z * [new tag] ciflow/periodic-rocm-mi200/169425 -> ciflow/periodic-rocm-mi200/169425 2025-12-04T09:20:54.5623197Z * [new tag] ciflow/periodic-rocm-mi300/166517 -> ciflow/periodic-rocm-mi300/166517 2025-12-04T09:20:54.5623377Z * [new tag] ciflow/periodic-rocm-mi300/169063 -> ciflow/periodic-rocm-mi300/169063 2025-12-04T09:20:54.5623784Z * [new tag] ciflow/periodic-rocm-mi300/169425 -> ciflow/periodic-rocm-mi300/169425 2025-12-04T09:20:54.5624384Z * [new tag] ciflow/periodic/054a2fd -> ciflow/periodic/054a2fd 2025-12-04T09:20:54.5624782Z * [new tag] ciflow/periodic/167207 -> ciflow/periodic/167207 2025-12-04T09:20:54.5628243Z * [new tag] ciflow/periodic/167978 -> ciflow/periodic/167978 2025-12-04T09:20:54.5628550Z * [new tag] ciflow/periodic/168096 -> ciflow/periodic/168096 2025-12-04T09:20:54.5628977Z * [new tag] ciflow/periodic/169286 -> ciflow/periodic/169286 2025-12-04T09:20:54.5629154Z * [new tag] ciflow/periodic/2a6d37d -> ciflow/periodic/2a6d37d 2025-12-04T09:20:54.5629375Z * [new tag] ciflow/periodic/317eeb8 -> ciflow/periodic/317eeb8 2025-12-04T09:20:54.5629577Z * [new tag] ciflow/periodic/3c32 -> ciflow/periodic/3c32 2025-12-04T09:20:54.5629724Z * [new tag] ciflow/periodic/3e98831 -> ciflow/periodic/3e98831 2025-12-04T09:20:54.5630551Z * [new tag] ciflow/periodic/7c648509a7470ace9fb2bae960dd4790f7e943e9 -> ciflow/periodic/7c648509a7470ace9fb2bae960dd4790f7e943e9 2025-12-04T09:20:54.5631517Z * [new tag] ciflow/periodic/94512-point -> ciflow/periodic/94512-point 2025-12-04T09:20:54.5632002Z * [new tag] ciflow/periodic/csl/test87519 -> ciflow/periodic/csl/test87519 2025-12-04T09:20:54.5634449Z * [new tag] ciflow/periodic/csltest88275 -> ciflow/periodic/csltest88275 2025-12-04T09:20:54.5634775Z * [new tag] ciflow/periodic/csltest88761 -> ciflow/periodic/csltest88761 2025-12-04T09:20:54.5634995Z * [new tag] ciflow/periodic/release_1.12 -> ciflow/periodic/release_1.12 2025-12-04T09:20:54.5635387Z * [new tag] ciflow/periodic/release_1.12.0 -> ciflow/periodic/release_1.12.0 2025-12-04T09:20:54.5635587Z * [new tag] ciflow/periodic/sha-ec5b83 -> ciflow/periodic/sha-ec5b83 2025-12-04T09:20:54.5635884Z * [new tag] ciflow/pull/167207 -> ciflow/pull/167207 2025-12-04T09:20:54.5636961Z * [new tag] ciflow/quantization-periodic/169207 -> ciflow/quantization-periodic/169207 2025-12-04T09:20:54.5637197Z * [new tag] ciflow/rocm-mi200/165545 -> ciflow/rocm-mi200/165545 2025-12-04T09:20:54.5637654Z * [new tag] ciflow/rocm-mi200/165997 -> ciflow/rocm-mi200/165997 2025-12-04T09:20:54.5638113Z * [new tag] ciflow/rocm-mi200/168096 -> ciflow/rocm-mi200/168096 2025-12-04T09:20:54.5640587Z * [new tag] ciflow/rocm-mi200/168275 -> ciflow/rocm-mi200/168275 2025-12-04T09:20:54.5640755Z * [new tag] ciflow/rocm-mi200/169063 -> ciflow/rocm-mi200/169063 2025-12-04T09:20:54.5640910Z * [new tag] ciflow/rocm-mi200/169356 -> ciflow/rocm-mi200/169356 2025-12-04T09:20:54.5641038Z * [new tag] ciflow/rocm-mi200/169425 -> ciflow/rocm-mi200/169425 2025-12-04T09:20:54.5641162Z * [new tag] ciflow/rocm-mi300/165545 -> ciflow/rocm-mi300/165545 2025-12-04T09:20:54.5641530Z * [new tag] ciflow/rocm-mi300/167157 -> ciflow/rocm-mi300/167157 2025-12-04T09:20:54.5642043Z * [new tag] ciflow/rocm-mi300/168096 -> ciflow/rocm-mi300/168096 2025-12-04T09:20:54.5642397Z * [new tag] ciflow/rocm-mi300/169063 -> ciflow/rocm-mi300/169063 2025-12-04T09:20:54.5642975Z * [new tag] ciflow/rocm-mi300/169425 -> ciflow/rocm-mi300/169425 2025-12-04T09:20:54.5643691Z * [new tag] ciflow/rocm-mi355/167157 -> ciflow/rocm-mi355/167157 2025-12-04T09:20:54.5643985Z * [new tag] ciflow/rocm-mi355/168275 -> ciflow/rocm-mi355/168275 2025-12-04T09:20:54.5644441Z * [new tag] ciflow/rocm-mi355/169425 -> ciflow/rocm-mi355/169425 2025-12-04T09:20:54.5647561Z * [new tag] ciflow/rocm-navi31/168275 -> ciflow/rocm-navi31/168275 2025-12-04T09:20:54.5647882Z * [new tag] ciflow/rocm-navi31/169425 -> ciflow/rocm-navi31/169425 2025-12-04T09:20:54.5648025Z * [new tag] ciflow/rocm/115316 -> ciflow/rocm/115316 2025-12-04T09:20:54.5648147Z * [new tag] ciflow/rocm/148492 -> ciflow/rocm/148492 2025-12-04T09:20:54.5648389Z * [new tag] ciflow/rocm/160685 -> ciflow/rocm/160685 2025-12-04T09:20:54.5648860Z * [new tag] ciflow/rocm/161607 -> ciflow/rocm/161607 2025-12-04T09:20:54.5649499Z * [new tag] ciflow/rocm/162052 -> ciflow/rocm/162052 2025-12-04T09:20:54.5649653Z * [new tag] ciflow/rocm/165997 -> ciflow/rocm/165997 2025-12-04T09:20:54.5650064Z * [new tag] ciflow/rocm/166165 -> ciflow/rocm/166165 2025-12-04T09:20:54.5650188Z * [new tag] ciflow/rocm/166517 -> ciflow/rocm/166517 2025-12-04T09:20:54.5650308Z * [new tag] ciflow/rocm/167207 -> ciflow/rocm/167207 2025-12-04T09:20:54.5650853Z * [new tag] ciflow/rocm/167536 -> ciflow/rocm/167536 2025-12-04T09:20:54.5651316Z * [new tag] ciflow/rocm/167781 -> ciflow/rocm/167781 2025-12-04T09:20:54.5652134Z * [new tag] ciflow/rocm/167989 -> ciflow/rocm/167989 2025-12-04T09:20:54.5652409Z * [new tag] ciflow/rocm/168073 -> ciflow/rocm/168073 2025-12-04T09:20:54.5654858Z * [new tag] ciflow/rocm/168195 -> ciflow/rocm/168195 2025-12-04T09:20:54.5655018Z * [new tag] ciflow/rocm/168939 -> ciflow/rocm/168939 2025-12-04T09:20:54.5655136Z * [new tag] ciflow/rocm/168971 -> ciflow/rocm/168971 2025-12-04T09:20:54.5655272Z * [new tag] ciflow/rocm/169024 -> ciflow/rocm/169024 2025-12-04T09:20:54.5655401Z * [new tag] ciflow/rocm/169200 -> ciflow/rocm/169200 2025-12-04T09:20:54.5655682Z * [new tag] ciflow/rocm/169216 -> ciflow/rocm/169216 2025-12-04T09:20:54.5655824Z * [new tag] ciflow/rocm/169312 -> ciflow/rocm/169312 2025-12-04T09:20:54.5656217Z * [new tag] ciflow/rocm/169380 -> ciflow/rocm/169380 2025-12-04T09:20:54.5657329Z * [new tag] ciflow/rocm/169427 -> ciflow/rocm/169427 2025-12-04T09:20:54.5657498Z * [new tag] ciflow/rocm/169455 -> ciflow/rocm/169455 2025-12-04T09:20:54.5657748Z * [new tag] ciflow/rocm/169470 -> ciflow/rocm/169470 2025-12-04T09:20:54.5658239Z * [new tag] ciflow/rocm/169471 -> ciflow/rocm/169471 2025-12-04T09:20:54.5658668Z * [new tag] ciflow/rocm/169472 -> ciflow/rocm/169472 2025-12-04T09:20:54.5659583Z * [new tag] ciflow/rocm/169514 -> ciflow/rocm/169514 2025-12-04T09:20:54.5660141Z * [new tag] ciflow/slow/01c7106 -> ciflow/slow/01c7106 2025-12-04T09:20:54.5660581Z * [new tag] ciflow/slow/0577043 -> ciflow/slow/0577043 2025-12-04T09:20:54.5661785Z * [new tag] ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym -> ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym 2025-12-04T09:20:54.5661915Z * [new tag] ciflow/slow/0e81104 -> ciflow/slow/0e81104 2025-12-04T09:20:54.5662287Z * [new tag] ciflow/slow/167207 -> ciflow/slow/167207 2025-12-04T09:20:54.5664153Z * [new tag] ciflow/slow/168050 -> ciflow/slow/168050 2025-12-04T09:20:54.5664345Z * [new tag] ciflow/slow/1732077 -> ciflow/slow/1732077 2025-12-04T09:20:54.5664520Z * [new tag] ciflow/slow/187eb7c -> ciflow/slow/187eb7c 2025-12-04T09:20:54.5665969Z * [new tag] ciflow/slow/1faef89 -> ciflow/slow/1faef89 2025-12-04T09:20:54.5666282Z * [new tag] ciflow/slow/3920ec1 -> ciflow/slow/3920ec1 2025-12-04T09:20:54.5666495Z * [new tag] ciflow/slow/3b7c6b2 -> ciflow/slow/3b7c6b2 2025-12-04T09:20:54.5666914Z * [new tag] ciflow/slow/59a3759 -> ciflow/slow/59a3759 2025-12-04T09:20:54.5667058Z * [new tag] ciflow/slow/70ef0bb -> ciflow/slow/70ef0bb 2025-12-04T09:20:54.5667714Z * [new tag] ciflow/slow/788ff06 -> ciflow/slow/788ff06 2025-12-04T09:20:54.5669318Z * [new tag] ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym -> ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym 2025-12-04T09:20:54.5669584Z * [new tag] ciflow/slow/9d85864 -> ciflow/slow/9d85864 2025-12-04T09:20:54.5669892Z * [new tag] ciflow/slow/9ffad5b -> ciflow/slow/9ffad5b 2025-12-04T09:20:54.5670218Z * [new tag] ciflow/slow/a206e8b -> ciflow/slow/a206e8b 2025-12-04T09:20:54.5671608Z * [new tag] ciflow/slow/a837609 -> ciflow/slow/a837609 2025-12-04T09:20:54.5671900Z * [new tag] ciflow/slow/af841f3 -> ciflow/slow/af841f3 2025-12-04T09:20:54.5672403Z * [new tag] ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym -> ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym 2025-12-04T09:20:54.5672897Z * [new tag] ciflow/torchbench/168175 -> ciflow/torchbench/168175 2025-12-04T09:20:54.5674809Z * [new tag] ciflow/trunk/148492 -> ciflow/trunk/148492 2025-12-04T09:20:54.5675100Z * [new tag] ciflow/trunk/157149 -> ciflow/trunk/157149 2025-12-04T09:20:54.5675443Z * [new tag] ciflow/trunk/157994 -> ciflow/trunk/157994 2025-12-04T09:20:54.5675567Z * [new tag] ciflow/trunk/159718 -> ciflow/trunk/159718 2025-12-04T09:20:54.5675704Z * [new tag] ciflow/trunk/160685 -> ciflow/trunk/160685 2025-12-04T09:20:54.5675824Z * [new tag] ciflow/trunk/160729 -> ciflow/trunk/160729 2025-12-04T09:20:54.5676304Z * [new tag] ciflow/trunk/162275 -> ciflow/trunk/162275 2025-12-04T09:20:54.5676758Z * [new tag] ciflow/trunk/162795 -> ciflow/trunk/162795 2025-12-04T09:20:54.5677234Z * [new tag] ciflow/trunk/163245 -> ciflow/trunk/163245 2025-12-04T09:20:54.5677687Z * [new tag] ciflow/trunk/163942 -> ciflow/trunk/163942 2025-12-04T09:20:54.5678246Z * [new tag] ciflow/trunk/165274 -> ciflow/trunk/165274 2025-12-04T09:20:54.5679077Z * [new tag] ciflow/trunk/165483 -> ciflow/trunk/165483 2025-12-04T09:20:54.5680087Z * [new tag] ciflow/trunk/165728 -> ciflow/trunk/165728 2025-12-04T09:20:54.5680646Z * [new tag] ciflow/trunk/165922 -> ciflow/trunk/165922 2025-12-04T09:20:54.5681065Z * [new tag] ciflow/trunk/166075 -> ciflow/trunk/166075 2025-12-04T09:20:54.5682552Z * [new tag] ciflow/trunk/166165 -> ciflow/trunk/166165 2025-12-04T09:20:54.5683024Z * [new tag] ciflow/trunk/166829 -> ciflow/trunk/166829 2025-12-04T09:20:54.5683470Z * [new tag] ciflow/trunk/166843 -> ciflow/trunk/166843 2025-12-04T09:20:54.5683940Z * [new tag] ciflow/trunk/166876 -> ciflow/trunk/166876 2025-12-04T09:20:54.5684403Z * [new tag] ciflow/trunk/167207 -> ciflow/trunk/167207 2025-12-04T09:20:54.5684814Z * [new tag] ciflow/trunk/167536 -> ciflow/trunk/167536 2025-12-04T09:20:54.5685245Z * [new tag] ciflow/trunk/167552 -> ciflow/trunk/167552 2025-12-04T09:20:54.5685679Z * [new tag] ciflow/trunk/167555 -> ciflow/trunk/167555 2025-12-04T09:20:54.5686114Z * [new tag] ciflow/trunk/167599 -> ciflow/trunk/167599 2025-12-04T09:20:54.5686838Z * [new tag] ciflow/trunk/167659 -> ciflow/trunk/167659 2025-12-04T09:20:54.5687247Z * [new tag] ciflow/trunk/167672 -> ciflow/trunk/167672 2025-12-04T09:20:54.5687577Z * [new tag] ciflow/trunk/167742 -> ciflow/trunk/167742 2025-12-04T09:20:54.5688080Z * [new tag] ciflow/trunk/167781 -> ciflow/trunk/167781 2025-12-04T09:20:54.5688409Z * [new tag] ciflow/trunk/167837 -> ciflow/trunk/167837 2025-12-04T09:20:54.5688721Z * [new tag] ciflow/trunk/167887 -> ciflow/trunk/167887 2025-12-04T09:20:54.5689500Z * [new tag] ciflow/trunk/167978 -> ciflow/trunk/167978 2025-12-04T09:20:54.5689908Z * [new tag] ciflow/trunk/168050 -> ciflow/trunk/168050 2025-12-04T09:20:54.5690245Z * [new tag] ciflow/trunk/168051 -> ciflow/trunk/168051 2025-12-04T09:20:54.5691466Z * [new tag] ciflow/trunk/168096 -> ciflow/trunk/168096 2025-12-04T09:20:54.5691797Z * [new tag] ciflow/trunk/168127 -> ciflow/trunk/168127 2025-12-04T09:20:54.5692273Z * [new tag] ciflow/trunk/168157 -> ciflow/trunk/168157 2025-12-04T09:20:54.5692698Z * [new tag] ciflow/trunk/168175 -> ciflow/trunk/168175 2025-12-04T09:20:54.5702596Z * [new tag] ciflow/trunk/168209 -> ciflow/trunk/168209 2025-12-04T09:20:54.5707648Z * [new tag] ciflow/trunk/168213 -> ciflow/trunk/168213 2025-12-04T09:20:54.5712648Z * [new tag] ciflow/trunk/168226 -> ciflow/trunk/168226 2025-12-04T09:20:54.5717731Z * [new tag] ciflow/trunk/168262 -> ciflow/trunk/168262 2025-12-04T09:20:54.5719010Z * [new tag] ciflow/trunk/168275 -> ciflow/trunk/168275 2025-12-04T09:20:54.5719435Z * [new tag] ciflow/trunk/168328 -> ciflow/trunk/168328 2025-12-04T09:20:54.5719902Z * [new tag] ciflow/trunk/168368 -> ciflow/trunk/168368 2025-12-04T09:20:54.5720258Z * [new tag] ciflow/trunk/168917 -> ciflow/trunk/168917 2025-12-04T09:20:54.5720579Z * [new tag] ciflow/trunk/168933 -> ciflow/trunk/168933 2025-12-04T09:20:54.5721276Z * [new tag] ciflow/trunk/168941 -> ciflow/trunk/168941 2025-12-04T09:20:54.5721644Z * [new tag] ciflow/trunk/168955 -> ciflow/trunk/168955 2025-12-04T09:20:54.5722304Z * [new tag] ciflow/trunk/168980 -> ciflow/trunk/168980 2025-12-04T09:20:54.5722644Z * [new tag] ciflow/trunk/169004 -> ciflow/trunk/169004 2025-12-04T09:20:54.5722946Z * [new tag] ciflow/trunk/169006 -> ciflow/trunk/169006 2025-12-04T09:20:54.5723243Z * [new tag] ciflow/trunk/169023 -> ciflow/trunk/169023 2025-12-04T09:20:54.5723553Z * [new tag] ciflow/trunk/169025 -> ciflow/trunk/169025 2025-12-04T09:20:54.5723864Z * [new tag] ciflow/trunk/169048 -> ciflow/trunk/169048 2025-12-04T09:20:54.5724176Z * [new tag] ciflow/trunk/169066 -> ciflow/trunk/169066 2025-12-04T09:20:54.5724466Z * [new tag] ciflow/trunk/169091 -> ciflow/trunk/169091 2025-12-04T09:20:54.5724775Z * [new tag] ciflow/trunk/169102 -> ciflow/trunk/169102 2025-12-04T09:20:54.5725084Z * [new tag] ciflow/trunk/169103 -> ciflow/trunk/169103 2025-12-04T09:20:54.5725395Z * [new tag] ciflow/trunk/169125 -> ciflow/trunk/169125 2025-12-04T09:20:54.5725703Z * [new tag] ciflow/trunk/169139 -> ciflow/trunk/169139 2025-12-04T09:20:54.5725996Z * [new tag] ciflow/trunk/169148 -> ciflow/trunk/169148 2025-12-04T09:20:54.5726307Z * [new tag] ciflow/trunk/169151 -> ciflow/trunk/169151 2025-12-04T09:20:54.5726610Z * [new tag] ciflow/trunk/169156 -> ciflow/trunk/169156 2025-12-04T09:20:54.5726926Z * [new tag] ciflow/trunk/169176 -> ciflow/trunk/169176 2025-12-04T09:20:54.5727234Z * [new tag] ciflow/trunk/169204 -> ciflow/trunk/169204 2025-12-04T09:20:54.5727684Z * [new tag] ciflow/trunk/169207 -> ciflow/trunk/169207 2025-12-04T09:20:54.5727992Z * [new tag] ciflow/trunk/169211 -> ciflow/trunk/169211 2025-12-04T09:20:54.5728289Z * [new tag] ciflow/trunk/169231 -> ciflow/trunk/169231 2025-12-04T09:20:54.5728652Z * [new tag] ciflow/trunk/169260 -> ciflow/trunk/169260 2025-12-04T09:20:54.5728950Z * [new tag] ciflow/trunk/169271 -> ciflow/trunk/169271 2025-12-04T09:20:54.5729257Z * [new tag] ciflow/trunk/169280 -> ciflow/trunk/169280 2025-12-04T09:20:54.5729549Z * [new tag] ciflow/trunk/169281 -> ciflow/trunk/169281 2025-12-04T09:20:54.5729842Z * [new tag] ciflow/trunk/169286 -> ciflow/trunk/169286 2025-12-04T09:20:54.5730304Z * [new tag] ciflow/trunk/169293 -> ciflow/trunk/169293 2025-12-04T09:20:54.5730619Z * [new tag] ciflow/trunk/169296 -> ciflow/trunk/169296 2025-12-04T09:20:54.5730925Z * [new tag] ciflow/trunk/169304 -> ciflow/trunk/169304 2025-12-04T09:20:54.5731229Z * [new tag] ciflow/trunk/169305 -> ciflow/trunk/169305 2025-12-04T09:20:54.5731525Z * [new tag] ciflow/trunk/169312 -> ciflow/trunk/169312 2025-12-04T09:20:54.5731829Z * [new tag] ciflow/trunk/169328 -> ciflow/trunk/169328 2025-12-04T09:20:54.5732130Z * [new tag] ciflow/trunk/169343 -> ciflow/trunk/169343 2025-12-04T09:20:54.5732432Z * [new tag] ciflow/trunk/169355 -> ciflow/trunk/169355 2025-12-04T09:20:54.5732730Z * [new tag] ciflow/trunk/169370 -> ciflow/trunk/169370 2025-12-04T09:20:54.5733020Z * [new tag] ciflow/trunk/169379 -> ciflow/trunk/169379 2025-12-04T09:20:54.5733308Z * [new tag] ciflow/trunk/169380 -> ciflow/trunk/169380 2025-12-04T09:20:54.5733601Z * [new tag] ciflow/trunk/169385 -> ciflow/trunk/169385 2025-12-04T09:20:54.5733883Z * [new tag] ciflow/trunk/169387 -> ciflow/trunk/169387 2025-12-04T09:20:54.5734175Z * [new tag] ciflow/trunk/169410 -> ciflow/trunk/169410 2025-12-04T09:20:54.5734470Z * [new tag] ciflow/trunk/169412 -> ciflow/trunk/169412 2025-12-04T09:20:54.5734754Z * [new tag] ciflow/trunk/169418 -> ciflow/trunk/169418 2025-12-04T09:20:54.5735041Z * [new tag] ciflow/trunk/169423 -> ciflow/trunk/169423 2025-12-04T09:20:54.5735329Z * [new tag] ciflow/trunk/169427 -> ciflow/trunk/169427 2025-12-04T09:20:54.5735618Z * [new tag] ciflow/trunk/169430 -> ciflow/trunk/169430 2025-12-04T09:20:54.5735907Z * [new tag] ciflow/trunk/169437 -> ciflow/trunk/169437 2025-12-04T09:20:54.5736200Z * [new tag] ciflow/trunk/169442 -> ciflow/trunk/169442 2025-12-04T09:20:54.5736494Z * [new tag] ciflow/trunk/169452 -> ciflow/trunk/169452 2025-12-04T09:20:54.5736780Z * [new tag] ciflow/trunk/169454 -> ciflow/trunk/169454 2025-12-04T09:20:54.5737066Z * [new tag] ciflow/trunk/169459 -> ciflow/trunk/169459 2025-12-04T09:20:54.5737353Z * [new tag] ciflow/trunk/169474 -> ciflow/trunk/169474 2025-12-04T09:20:54.5737643Z * [new tag] ciflow/trunk/169475 -> ciflow/trunk/169475 2025-12-04T09:20:54.5737926Z * [new tag] ciflow/trunk/169476 -> ciflow/trunk/169476 2025-12-04T09:20:54.5738213Z * [new tag] ciflow/trunk/169487 -> ciflow/trunk/169487 2025-12-04T09:20:54.5738500Z * [new tag] ciflow/trunk/169497 -> ciflow/trunk/169497 2025-12-04T09:20:54.5738912Z * [new tag] ciflow/trunk/169503 -> ciflow/trunk/169503 2025-12-04T09:20:54.5739200Z * [new tag] ciflow/trunk/169505 -> ciflow/trunk/169505 2025-12-04T09:20:54.5739492Z * [new tag] ciflow/trunk/169507 -> ciflow/trunk/169507 2025-12-04T09:20:54.5739787Z * [new tag] ciflow/trunk/169514 -> ciflow/trunk/169514 2025-12-04T09:20:54.5740130Z * [new tag] ciflow/trunk/169517 -> ciflow/trunk/169517 2025-12-04T09:20:54.5740426Z * [new tag] ciflow/trunk/169519 -> ciflow/trunk/169519 2025-12-04T09:20:54.5740717Z * [new tag] ciflow/trunk/169528 -> ciflow/trunk/169528 2025-12-04T09:20:54.5741007Z * [new tag] ciflow/trunk/169541 -> ciflow/trunk/169541 2025-12-04T09:20:54.5741290Z * [new tag] ciflow/trunk/169555 -> ciflow/trunk/169555 2025-12-04T09:20:54.5741582Z * [new tag] ciflow/unstable/123 -> ciflow/unstable/123 2025-12-04T09:20:54.5741880Z * [new tag] ciflow/vllm/165270 -> ciflow/vllm/165270 2025-12-04T09:20:54.5742180Z * [new tag] ciflow/vllm/165274 -> ciflow/vllm/165274 2025-12-04T09:20:54.5742479Z * [new tag] ciflow/vllm/166494 -> ciflow/vllm/166494 2025-12-04T09:20:54.5742773Z * [new tag] ciflow/vllm/169219 -> ciflow/vllm/169219 2025-12-04T09:20:54.5743060Z * [new tag] ciflow/vllm/169220 -> ciflow/vllm/169220 2025-12-04T09:20:54.5743354Z * [new tag] ciflow/xpu/157994 -> ciflow/xpu/157994 2025-12-04T09:20:54.5743638Z * [new tag] ciflow/xpu/159718 -> ciflow/xpu/159718 2025-12-04T09:20:54.5743927Z * [new tag] ciflow/xpu/161940 -> ciflow/xpu/161940 2025-12-04T09:20:54.5744214Z * [new tag] ciflow/xpu/163251 -> ciflow/xpu/163251 2025-12-04T09:20:54.5744494Z * [new tag] ciflow/xpu/166829 -> ciflow/xpu/166829 2025-12-04T09:20:54.5744780Z * [new tag] ciflow/xpu/166843 -> ciflow/xpu/166843 2025-12-04T09:20:54.5745064Z * [new tag] ciflow/xpu/167972 -> ciflow/xpu/167972 2025-12-04T09:20:54.5745346Z * [new tag] ciflow/xpu/167981 -> ciflow/xpu/167981 2025-12-04T09:20:54.5745626Z * [new tag] ciflow/xpu/168213 -> ciflow/xpu/168213 2025-12-04T09:20:54.5745909Z * [new tag] ciflow/xpu/168262 -> ciflow/xpu/168262 2025-12-04T09:20:54.5746193Z * [new tag] ciflow/xpu/168328 -> ciflow/xpu/168328 2025-12-04T09:20:54.5746475Z * [new tag] ciflow/xpu/168950 -> ciflow/xpu/168950 2025-12-04T09:20:54.5746754Z * [new tag] ciflow/xpu/169039 -> ciflow/xpu/169039 2025-12-04T09:20:54.5747042Z * [new tag] ciflow/xpu/169200 -> ciflow/xpu/169200 2025-12-04T09:20:54.5747331Z * [new tag] ciflow/xpu/169203 -> ciflow/xpu/169203 2025-12-04T09:20:54.5747610Z * [new tag] ciflow/xpu/169230 -> ciflow/xpu/169230 2025-12-04T09:20:54.5747891Z * [new tag] ciflow/xpu/169231 -> ciflow/xpu/169231 2025-12-04T09:20:54.5748178Z * [new tag] ciflow/xpu/169241 -> ciflow/xpu/169241 2025-12-04T09:20:54.5748462Z * [new tag] ciflow/xpu/169280 -> ciflow/xpu/169280 2025-12-04T09:20:54.5748742Z * [new tag] ciflow/xpu/169296 -> ciflow/xpu/169296 2025-12-04T09:20:54.5749035Z * [new tag] ciflow/xpu/169353 -> ciflow/xpu/169353 2025-12-04T09:20:54.5749329Z * [new tag] ciflow/xpu/169410 -> ciflow/xpu/169410 2025-12-04T09:20:54.5749620Z * [new tag] ciflow/xpu/169442 -> ciflow/xpu/169442 2025-12-04T09:20:54.5749904Z * [new tag] ciflow/xpu/169555 -> ciflow/xpu/169555 2025-12-04T09:20:54.5750271Z * [new tag] cslpull75 -> cslpull75 2025-12-04T09:20:54.5750549Z * [new tag] cslpull76 -> cslpull76 2025-12-04T09:20:54.5750822Z * [new tag] cslpull77 -> cslpull77 2025-12-04T09:20:54.5751131Z * [new tag] cslpull78 -> cslpull78 2025-12-04T09:20:54.5751408Z * [new tag] cslpull79 -> cslpull79 2025-12-04T09:20:54.5751681Z * [new tag] cslpull80 -> cslpull80 2025-12-04T09:20:54.5751948Z * [new tag] cslpull81 -> cslpull81 2025-12-04T09:20:54.5752217Z * [new tag] cslpull82 -> cslpull82 2025-12-04T09:20:54.5752485Z * [new tag] cslpull83 -> cslpull83 2025-12-04T09:20:54.5752747Z * [new tag] cslpull84 -> cslpull84 2025-12-04T09:20:54.5753020Z * [new tag] cslpull85 -> cslpull85 2025-12-04T09:20:54.5753291Z * [new tag] cslpull86 -> cslpull86 2025-12-04T09:20:54.5753560Z * [new tag] cslpull87 -> cslpull87 2025-12-04T09:20:54.5753821Z * [new tag] cslpull88 -> cslpull88 2025-12-04T09:20:54.5754093Z * [new tag] cslpull89 -> cslpull89 2025-12-04T09:20:54.5754364Z * [new tag] cslpull90 -> cslpull90 2025-12-04T09:20:54.5754633Z * [new tag] cslpull91 -> cslpull91 2025-12-04T09:20:54.5755117Z * [new tag] cslpull92 -> cslpull92 2025-12-04T09:20:54.5755497Z * [new tag] flight_5 -> flight_5 2025-12-04T09:20:54.5755874Z * [new tag] flight_5.1 -> flight_5.1 2025-12-04T09:20:54.5756245Z * [new tag] flight_5.2 -> flight_5.2 2025-12-04T09:20:54.5756982Z * [new tag] flight_5.3 -> flight_5.3 2025-12-04T09:20:54.5757403Z * [new tag] forpull1 -> forpull1 2025-12-04T09:20:54.5757721Z * [new tag] malfet/tag-2ef5611 -> malfet/tag-2ef5611 2025-12-04T09:20:54.5758060Z * [new tag] malfet/tag-317b1a0 -> malfet/tag-317b1a0 2025-12-04T09:20:54.5759299Z * [new tag] malfet/tag-ec6f767 -> malfet/tag-ec6f767 2025-12-04T09:20:54.5759763Z * [new tag] nightly-binary -> nightly-binary 2025-12-04T09:20:54.5760185Z * [new tag] sqzhang_flight4_plus -> sqzhang_flight4_plus 2025-12-04T09:20:54.5760595Z * [new tag] sqzhang_flight_3 -> sqzhang_flight_3 2025-12-04T09:20:54.5761555Z * [new tag] trunk/02d8bd6974cf84b721680d773dbdb1b6f40ce272 -> trunk/02d8bd6974cf84b721680d773dbdb1b6f40ce272 2025-12-04T09:20:54.5762540Z * [new tag] trunk/066997fb38ade71e00d78e9d572e380b5f02bd3e -> trunk/066997fb38ade71e00d78e9d572e380b5f02bd3e 2025-12-04T09:20:54.5763117Z * [new tag] trunk/076e7b19fa1d481ad778d06d2b49ba57d3ce8c88 -> trunk/076e7b19fa1d481ad778d06d2b49ba57d3ce8c88 2025-12-04T09:20:54.5763687Z * [new tag] trunk/07dcc0b83db3211653a38565a24e15acdba75654 -> trunk/07dcc0b83db3211653a38565a24e15acdba75654 2025-12-04T09:20:54.5764252Z * [new tag] trunk/082e96b68dfcd16cab7cfafc4d3d055767dab3eb -> trunk/082e96b68dfcd16cab7cfafc4d3d055767dab3eb 2025-12-04T09:20:54.5764830Z * [new tag] trunk/088048f2fea28ff7d450f65c72419ca45780d30b -> trunk/088048f2fea28ff7d450f65c72419ca45780d30b 2025-12-04T09:20:54.5765429Z * [new tag] trunk/09076941a95c76f4d9ad189d064dfd8baa39e672 -> trunk/09076941a95c76f4d9ad189d064dfd8baa39e672 2025-12-04T09:20:54.5766841Z * [new tag] trunk/0b80a4c62b94402844bf221791c096b0035c6d75 -> trunk/0b80a4c62b94402844bf221791c096b0035c6d75 2025-12-04T09:20:54.5767417Z * [new tag] trunk/0bbbdf1750567a980634ad907a325357ba8ba8f2 -> trunk/0bbbdf1750567a980634ad907a325357ba8ba8f2 2025-12-04T09:20:54.5768053Z * [new tag] trunk/0c281dd78773b2bc17c58ead0e4cd4ac46e775c5 -> trunk/0c281dd78773b2bc17c58ead0e4cd4ac46e775c5 2025-12-04T09:20:54.5768619Z * [new tag] trunk/135f3753c418a6879b1954904184937b67e61688 -> trunk/135f3753c418a6879b1954904184937b67e61688 2025-12-04T09:20:54.5769165Z * [new tag] trunk/15da21026cb13cd20257dc9e96830db108743c10 -> trunk/15da21026cb13cd20257dc9e96830db108743c10 2025-12-04T09:20:54.5769732Z * [new tag] trunk/166efdad2ac827f30fb02504c6017520257f88ec -> trunk/166efdad2ac827f30fb02504c6017520257f88ec 2025-12-04T09:20:54.5770288Z * [new tag] trunk/174272c15fae553d8488140af931f7d8050a313f -> trunk/174272c15fae553d8488140af931f7d8050a313f 2025-12-04T09:20:54.5770869Z * [new tag] trunk/18f3ca08f13b8de61307f5e8cd7d4cccb67e9d11 -> trunk/18f3ca08f13b8de61307f5e8cd7d4cccb67e9d11 2025-12-04T09:20:54.5771603Z * [new tag] trunk/1902eddfe655a15ebcf2c72bd81ade110fdeef63 -> trunk/1902eddfe655a15ebcf2c72bd81ade110fdeef63 2025-12-04T09:20:54.5772311Z * [new tag] trunk/195f92e98d3d66738577f11f22c4b5c8a1c76dd5 -> trunk/195f92e98d3d66738577f11f22c4b5c8a1c76dd5 2025-12-04T09:20:54.5773005Z * [new tag] trunk/1aa13e17de39e3c768ea7aebaad166ce72a06676 -> trunk/1aa13e17de39e3c768ea7aebaad166ce72a06676 2025-12-04T09:20:54.5773581Z * [new tag] trunk/1afe2832f58e24e54a5bfda5a5afa9b96fdea40e -> trunk/1afe2832f58e24e54a5bfda5a5afa9b96fdea40e 2025-12-04T09:20:54.5774157Z * [new tag] trunk/1c87554d74140eaee964ca8b1832cede67f5f520 -> trunk/1c87554d74140eaee964ca8b1832cede67f5f520 2025-12-04T09:20:54.5774711Z * [new tag] trunk/1ccb743b7b5be955f49736c162c4f5004b8a0dd8 -> trunk/1ccb743b7b5be955f49736c162c4f5004b8a0dd8 2025-12-04T09:20:54.5775267Z * [new tag] trunk/1cee47d6ce0a02227185b566593f002dd639ca0c -> trunk/1cee47d6ce0a02227185b566593f002dd639ca0c 2025-12-04T09:20:54.5775825Z * [new tag] trunk/1d21b4df2babe322e5d085ceb6de884eb260a62d -> trunk/1d21b4df2babe322e5d085ceb6de884eb260a62d 2025-12-04T09:20:54.5776395Z * [new tag] trunk/1e34fb2550e4aa650314f7a6d9f6daf4da7478a8 -> trunk/1e34fb2550e4aa650314f7a6d9f6daf4da7478a8 2025-12-04T09:20:54.5776951Z * [new tag] trunk/1e526fb5b1d93bfc70691c5c3955fdffc1b7b7de -> trunk/1e526fb5b1d93bfc70691c5c3955fdffc1b7b7de 2025-12-04T09:20:54.5777520Z * [new tag] trunk/1ee32a8b1f554a312d79bad01ded24f38cd95543 -> trunk/1ee32a8b1f554a312d79bad01ded24f38cd95543 2025-12-04T09:20:54.5778078Z * [new tag] trunk/201e2c4117eb9744594dad6a5c18213d7b4705d7 -> trunk/201e2c4117eb9744594dad6a5c18213d7b4705d7 2025-12-04T09:20:54.5778620Z * [new tag] trunk/2353a0f60eb4b4cb6675907a7fa9fbedc1c02e7f -> trunk/2353a0f60eb4b4cb6675907a7fa9fbedc1c02e7f 2025-12-04T09:20:54.5779161Z * [new tag] trunk/285779b1621cf9f073a062b0889a642d200308d9 -> trunk/285779b1621cf9f073a062b0889a642d200308d9 2025-12-04T09:20:54.5779708Z * [new tag] trunk/2887faaec6295d081580d09fce161201826c6d87 -> trunk/2887faaec6295d081580d09fce161201826c6d87 2025-12-04T09:20:54.5780250Z * [new tag] trunk/296e67c92635443c67b11c0ae1bd045f03ebb7bc -> trunk/296e67c92635443c67b11c0ae1bd045f03ebb7bc 2025-12-04T09:20:54.5780783Z * [new tag] trunk/29856679769b3dede478767e2fe6cfb51197cb25 -> trunk/29856679769b3dede478767e2fe6cfb51197cb25 2025-12-04T09:20:54.5781470Z * [new tag] trunk/29e5455a4740c326ab187c7aa7b5ef98034ea563 -> trunk/29e5455a4740c326ab187c7aa7b5ef98034ea563 2025-12-04T09:20:54.5782506Z * [new tag] trunk/2ac3ef882afb23136adc188975f0a8802fc68adf -> trunk/2ac3ef882afb23136adc188975f0a8802fc68adf 2025-12-04T09:20:54.5783364Z * [new tag] trunk/2bec68e73b64715354af076ad309335f943e36cd -> trunk/2bec68e73b64715354af076ad309335f943e36cd 2025-12-04T09:20:54.5783936Z * [new tag] trunk/2c87367e6f88662cd5cedbd1537748b7948c38e1 -> trunk/2c87367e6f88662cd5cedbd1537748b7948c38e1 2025-12-04T09:20:54.5784572Z * [new tag] trunk/2d1f78fe3ec13820f136a2e0336da12a25f41708 -> trunk/2d1f78fe3ec13820f136a2e0336da12a25f41708 2025-12-04T09:20:54.5785144Z * [new tag] trunk/2df6058f116a65722a0e03073402feb242572d35 -> trunk/2df6058f116a65722a0e03073402feb242572d35 2025-12-04T09:20:54.5785707Z * [new tag] trunk/2e0c2e170fe658c440775c8e5c44228aafcc47ec -> trunk/2e0c2e170fe658c440775c8e5c44228aafcc47ec 2025-12-04T09:20:54.5786288Z * [new tag] trunk/2f9b7dad7b5419b063bd0f2e204de192720ebb94 -> trunk/2f9b7dad7b5419b063bd0f2e204de192720ebb94 2025-12-04T09:20:54.5786882Z * [new tag] trunk/305168768a95d69c444df5cd334bb774edfe06f1 -> trunk/305168768a95d69c444df5cd334bb774edfe06f1 2025-12-04T09:20:54.5787493Z * [new tag] trunk/31fc12773026e8e00f054dd79ad9b2491e693b48 -> trunk/31fc12773026e8e00f054dd79ad9b2491e693b48 2025-12-04T09:20:54.5788078Z * [new tag] trunk/320de0c6b0a3e7c6d2693ea5c28d5d0156ba7991 -> trunk/320de0c6b0a3e7c6d2693ea5c28d5d0156ba7991 2025-12-04T09:20:54.5788643Z * [new tag] trunk/3418bd29475dff06695045fcdf93e7d0dac67da8 -> trunk/3418bd29475dff06695045fcdf93e7d0dac67da8 2025-12-04T09:20:54.5789199Z * [new tag] trunk/34a98608afa0cb5b48f0d6d30432fdd0a2614ddf -> trunk/34a98608afa0cb5b48f0d6d30432fdd0a2614ddf 2025-12-04T09:20:54.5789790Z * [new tag] trunk/35b7a9a26c5923d98aebaa41a031dae21788a9ee -> trunk/35b7a9a26c5923d98aebaa41a031dae21788a9ee 2025-12-04T09:20:54.5790355Z * [new tag] trunk/39d07dbf03a911bdd45d1af78d8638dc92074938 -> trunk/39d07dbf03a911bdd45d1af78d8638dc92074938 2025-12-04T09:20:54.5790916Z * [new tag] trunk/3cd98b4205ada151042cc7ff097a82d4a4b18725 -> trunk/3cd98b4205ada151042cc7ff097a82d4a4b18725 2025-12-04T09:20:54.5791479Z * [new tag] trunk/3d35fd20a78ff4d016fa80f4e5fad37191d7bcae -> trunk/3d35fd20a78ff4d016fa80f4e5fad37191d7bcae 2025-12-04T09:20:54.5792052Z * [new tag] trunk/409a5fee945c46a3edaf5df162812f201bfd7b2f -> trunk/409a5fee945c46a3edaf5df162812f201bfd7b2f 2025-12-04T09:20:54.5792621Z * [new tag] trunk/42e9005cda22da3f1c559c3649218cebd671027c -> trunk/42e9005cda22da3f1c559c3649218cebd671027c 2025-12-04T09:20:54.5793178Z * [new tag] trunk/43b94713bbf340d3c124fde02d0f73add4021247 -> trunk/43b94713bbf340d3c124fde02d0f73add4021247 2025-12-04T09:20:54.5793740Z * [new tag] trunk/44ac69388a4a5eb463dbd2a13f00d1e3b924566c -> trunk/44ac69388a4a5eb463dbd2a13f00d1e3b924566c 2025-12-04T09:20:54.5794316Z * [new tag] trunk/45d14e2497292be06ad36eaa1aaaf7c630a2586a -> trunk/45d14e2497292be06ad36eaa1aaaf7c630a2586a 2025-12-04T09:20:54.5794895Z * [new tag] trunk/45d310ad84854dff730c0b12e577d7998d978686 -> trunk/45d310ad84854dff730c0b12e577d7998d978686 2025-12-04T09:20:54.5795459Z * [new tag] trunk/47b28ddf7bd74b50fa93b307a7d3b183a6d77f54 -> trunk/47b28ddf7bd74b50fa93b307a7d3b183a6d77f54 2025-12-04T09:20:54.5796029Z * [new tag] trunk/481e5ab336275bd3acd5fa8a611b05b4469012af -> trunk/481e5ab336275bd3acd5fa8a611b05b4469012af 2025-12-04T09:20:54.5796584Z * [new tag] trunk/491731647f6b8a9345dcfb3bc9416aea254a7d96 -> trunk/491731647f6b8a9345dcfb3bc9416aea254a7d96 2025-12-04T09:20:54.5797144Z * [new tag] trunk/49a04d26088acc17d948ddd66920f3e16371e873 -> trunk/49a04d26088acc17d948ddd66920f3e16371e873 2025-12-04T09:20:54.5797962Z * [new tag] trunk/4bebc827c47d2f1f0fa1a417a5201a97aef3d985 -> trunk/4bebc827c47d2f1f0fa1a417a5201a97aef3d985 2025-12-04T09:20:54.5802218Z * [new tag] trunk/4c246677784c6a14bc2dbb9ff8773ef0a3a3222f -> trunk/4c246677784c6a14bc2dbb9ff8773ef0a3a3222f 2025-12-04T09:20:54.5802791Z * [new tag] trunk/4cfb47ff548b6d996641058cf04a70e311a4c3aa -> trunk/4cfb47ff548b6d996641058cf04a70e311a4c3aa 2025-12-04T09:20:54.5809142Z * [new tag] trunk/4e0061c1aa52f606dda8cfab0bd7591e588faf2c -> trunk/4e0061c1aa52f606dda8cfab0bd7591e588faf2c 2025-12-04T09:20:54.5810551Z * [new tag] trunk/4fefb8e7e942386ffac764a41b232241f82bea3a -> trunk/4fefb8e7e942386ffac764a41b232241f82bea3a 2025-12-04T09:20:54.5811140Z * [new tag] trunk/503b2640023521f5a35cd9a52fc8033d73a95d0d -> trunk/503b2640023521f5a35cd9a52fc8033d73a95d0d 2025-12-04T09:20:54.5811702Z * [new tag] trunk/518c2b1b3dab9a2ef2849e04b3bc2f20c1c41db9 -> trunk/518c2b1b3dab9a2ef2849e04b3bc2f20c1c41db9 2025-12-04T09:20:54.5812268Z * [new tag] trunk/5191b2fa68ba19960912bfd7fd721c79d76bb1f3 -> trunk/5191b2fa68ba19960912bfd7fd721c79d76bb1f3 2025-12-04T09:20:54.5812848Z * [new tag] trunk/52ac0f0dc4acacd219f1317fbc28ec631c01e07a -> trunk/52ac0f0dc4acacd219f1317fbc28ec631c01e07a 2025-12-04T09:20:54.5813412Z * [new tag] trunk/539ba711b029de9f191070f4f0d12f18f5b7f292 -> trunk/539ba711b029de9f191070f4f0d12f18f5b7f292 2025-12-04T09:20:54.5813966Z * [new tag] trunk/556375b55deebebbc56cb7aef81f4d52f031ba28 -> trunk/556375b55deebebbc56cb7aef81f4d52f031ba28 2025-12-04T09:20:54.5814499Z * [new tag] trunk/55c4ab554845481d0a69a3811937575fe8bb1a66 -> trunk/55c4ab554845481d0a69a3811937575fe8bb1a66 2025-12-04T09:20:54.5815018Z * [new tag] trunk/5634469fda9e5d98869c82c7d03bb08914245f96 -> trunk/5634469fda9e5d98869c82c7d03bb08914245f96 2025-12-04T09:20:54.5815541Z * [new tag] trunk/5778f6ff894686a975a9a23645178ae4c87ad5dc -> trunk/5778f6ff894686a975a9a23645178ae4c87ad5dc 2025-12-04T09:20:54.5816085Z * [new tag] trunk/587d63a3e07de5dc91065f9ef70bcacda9989068 -> trunk/587d63a3e07de5dc91065f9ef70bcacda9989068 2025-12-04T09:20:54.5816618Z * [new tag] trunk/597930f6b568852356ca9795dac76f9e4653adbd -> trunk/597930f6b568852356ca9795dac76f9e4653adbd 2025-12-04T09:20:54.5817147Z * [new tag] trunk/597df3a4e2a67b9fdbe1a89b2f4d74f822274db6 -> trunk/597df3a4e2a67b9fdbe1a89b2f4d74f822274db6 2025-12-04T09:20:54.5817697Z * [new tag] trunk/59abd50e931f4efb21b053f7a2911f5d8a49d883 -> trunk/59abd50e931f4efb21b053f7a2911f5d8a49d883 2025-12-04T09:20:54.5818237Z * [new tag] trunk/5a607febc04c3a2b5824c75f3f60307867439a2c -> trunk/5a607febc04c3a2b5824c75f3f60307867439a2c 2025-12-04T09:20:54.5818783Z * [new tag] trunk/5bf1cdf4755c54ef462b44cb8041b0a57311556b -> trunk/5bf1cdf4755c54ef462b44cb8041b0a57311556b 2025-12-04T09:20:54.5819313Z * [new tag] trunk/5f0030ba63d334d7e8c93a09e41403b89e4c573c -> trunk/5f0030ba63d334d7e8c93a09e41403b89e4c573c 2025-12-04T09:20:54.5819849Z * [new tag] trunk/5f21d27e71268464d362a96c9ac09ea475f7f202 -> trunk/5f21d27e71268464d362a96c9ac09ea475f7f202 2025-12-04T09:20:54.5820380Z * [new tag] trunk/5fafc13038c9988d9ac21fa793fbd5890604b447 -> trunk/5fafc13038c9988d9ac21fa793fbd5890604b447 2025-12-04T09:20:54.5820923Z * [new tag] trunk/61be54a31dc09b59d99b62176fb935aee0b924ef -> trunk/61be54a31dc09b59d99b62176fb935aee0b924ef 2025-12-04T09:20:54.5821452Z * [new tag] trunk/62d3ccd71484ed6a760d909b41487101bbc65719 -> trunk/62d3ccd71484ed6a760d909b41487101bbc65719 2025-12-04T09:20:54.5821983Z * [new tag] trunk/641cdb68ae27668eb441d0e49c87a0602c120c2b -> trunk/641cdb68ae27668eb441d0e49c87a0602c120c2b 2025-12-04T09:20:54.5822514Z * [new tag] trunk/65c4620d6bb0c6029f69762c22b91dda2294da9a -> trunk/65c4620d6bb0c6029f69762c22b91dda2294da9a 2025-12-04T09:20:54.5823041Z * [new tag] trunk/66004b993744b4106bf8afaba71f3c228a804206 -> trunk/66004b993744b4106bf8afaba71f3c228a804206 2025-12-04T09:20:54.5823625Z * [new tag] trunk/6658a04c7ca67acb64512341342e7b3ee13ee386 -> trunk/6658a04c7ca67acb64512341342e7b3ee13ee386 2025-12-04T09:20:54.5824170Z * [new tag] trunk/6864e309092a71f8ab0ca6a4dc7f8a4073fd31c4 -> trunk/6864e309092a71f8ab0ca6a4dc7f8a4073fd31c4 2025-12-04T09:20:54.5824812Z * [new tag] trunk/6c261c6cb07892c90ca19ed51c9705b1659a3f7d -> trunk/6c261c6cb07892c90ca19ed51c9705b1659a3f7d 2025-12-04T09:20:54.5825365Z * [new tag] trunk/6c8b6a043f1628188b6396b3a2a6e000ca68362b -> trunk/6c8b6a043f1628188b6396b3a2a6e000ca68362b 2025-12-04T09:20:54.5825905Z * [new tag] trunk/6ceb4a32f92ae67ce5d7d97931d17401ebf5ffa5 -> trunk/6ceb4a32f92ae67ce5d7d97931d17401ebf5ffa5 2025-12-04T09:20:54.5826437Z * [new tag] trunk/6e404e9b7d6f5fb0de86aa73888c3038248c17f8 -> trunk/6e404e9b7d6f5fb0de86aa73888c3038248c17f8 2025-12-04T09:20:54.5826987Z * [new tag] trunk/6ec30b490aee1db6bcdc7340abddef25784f08ec -> trunk/6ec30b490aee1db6bcdc7340abddef25784f08ec 2025-12-04T09:20:54.5827504Z * [new tag] trunk/6f2783a6c08e1db34275ff25176ffe9aebc30a71 -> trunk/6f2783a6c08e1db34275ff25176ffe9aebc30a71 2025-12-04T09:20:54.5828026Z * [new tag] trunk/6f53fefeb90ad3281119b5cfc4aa9ffd8a066e3d -> trunk/6f53fefeb90ad3281119b5cfc4aa9ffd8a066e3d 2025-12-04T09:20:54.5828569Z * [new tag] trunk/6f7dcf51e46d0c880db1a2f5c70de57adb576f4a -> trunk/6f7dcf51e46d0c880db1a2f5c70de57adb576f4a 2025-12-04T09:20:54.5829120Z * [new tag] trunk/6ff831180d2fa436c7f1c1af3adac641fce9d60e -> trunk/6ff831180d2fa436c7f1c1af3adac641fce9d60e 2025-12-04T09:20:54.5829663Z * [new tag] trunk/70076464a63ab218a7ceefb0e76ccd7131deb8f8 -> trunk/70076464a63ab218a7ceefb0e76ccd7131deb8f8 2025-12-04T09:20:54.5830360Z * [new tag] trunk/70d797a5fc109b20a517646fcaa819477cd0d485 -> trunk/70d797a5fc109b20a517646fcaa819477cd0d485 2025-12-04T09:20:54.5830898Z * [new tag] trunk/7348cb355ff0a6f79cd4871215aea72185748734 -> trunk/7348cb355ff0a6f79cd4871215aea72185748734 2025-12-04T09:20:54.5831432Z * [new tag] trunk/74fe26a1ebe32931783569f2e762e3c2c974901f -> trunk/74fe26a1ebe32931783569f2e762e3c2c974901f 2025-12-04T09:20:54.5831978Z * [new tag] trunk/76aeb8c7e0f795b3fddca134cbea9a69da3ee696 -> trunk/76aeb8c7e0f795b3fddca134cbea9a69da3ee696 2025-12-04T09:20:54.5832529Z * [new tag] trunk/7716da9fb23f27a65b41f9f016a2afadf281c18f -> trunk/7716da9fb23f27a65b41f9f016a2afadf281c18f 2025-12-04T09:20:54.5833064Z * [new tag] trunk/7741edd4ed665f3988052e260863efb508d61a03 -> trunk/7741edd4ed665f3988052e260863efb508d61a03 2025-12-04T09:20:54.5833595Z * [new tag] trunk/78adb3b3df41b45d2368b67226d2f864b78939a6 -> trunk/78adb3b3df41b45d2368b67226d2f864b78939a6 2025-12-04T09:20:54.5834136Z * [new tag] trunk/79d7b178225e5ed24d4e1db74e5abbff848f5fb7 -> trunk/79d7b178225e5ed24d4e1db74e5abbff848f5fb7 2025-12-04T09:20:54.5834672Z * [new tag] trunk/7a1e316115fc6996b3f2336822ba5d5f6179f0c3 -> trunk/7a1e316115fc6996b3f2336822ba5d5f6179f0c3 2025-12-04T09:20:54.5835213Z * [new tag] trunk/7a41b66367c38d0af3e8a90f7be48d6b281e7bca -> trunk/7a41b66367c38d0af3e8a90f7be48d6b281e7bca 2025-12-04T09:20:54.5835741Z * [new tag] trunk/7b7af390ea8541c611d1ce2018a6934188fc197b -> trunk/7b7af390ea8541c611d1ce2018a6934188fc197b 2025-12-04T09:20:54.5836268Z * [new tag] trunk/7ba4680f3755a560af81aa0f688791e367aa3609 -> trunk/7ba4680f3755a560af81aa0f688791e367aa3609 2025-12-04T09:20:54.5836811Z * [new tag] trunk/7bc2a66ded06a0b2549aa51d807edc5dc3e73d1b -> trunk/7bc2a66ded06a0b2549aa51d807edc5dc3e73d1b 2025-12-04T09:20:54.5837366Z * [new tag] trunk/7c648509a7470ace9fb2bae960dd4790f7e943e9 -> trunk/7c648509a7470ace9fb2bae960dd4790f7e943e9 2025-12-04T09:20:54.5838025Z * [new tag] trunk/7cbc2d034cecd21ab5c9707d0a9c525c17143fb8 -> trunk/7cbc2d034cecd21ab5c9707d0a9c525c17143fb8 2025-12-04T09:20:54.5838587Z * [new tag] trunk/7d1bbaf4ba301ea3fba6f3c7bc02d58f6417aaed -> trunk/7d1bbaf4ba301ea3fba6f3c7bc02d58f6417aaed 2025-12-04T09:20:54.5839228Z * [new tag] trunk/7d2a33e4ebf60b217a3cd77feae19231eb996fc8 -> trunk/7d2a33e4ebf60b217a3cd77feae19231eb996fc8 2025-12-04T09:20:54.5839766Z * [new tag] trunk/7eb625920054b1126a7d2d99818aaa188c6ba95e -> trunk/7eb625920054b1126a7d2d99818aaa188c6ba95e 2025-12-04T09:20:54.5840295Z * [new tag] trunk/7f55ba19c456a3d6cc443dd9edb6bb7cca677ead -> trunk/7f55ba19c456a3d6cc443dd9edb6bb7cca677ead 2025-12-04T09:20:54.5840833Z * [new tag] trunk/81af382128efa094d8702e18f2c133760904c718 -> trunk/81af382128efa094d8702e18f2c133760904c718 2025-12-04T09:20:54.5841369Z * [new tag] trunk/84149583d483e9c973c9a0feda70e4f3964947b0 -> trunk/84149583d483e9c973c9a0feda70e4f3964947b0 2025-12-04T09:20:54.5842171Z * [new tag] trunk/85a315917efe82c24306be805c584ec044951c75 -> trunk/85a315917efe82c24306be805c584ec044951c75 2025-12-04T09:20:54.5842747Z * [new tag] trunk/87329491c82a5f8c1cc4ec11d8f55a5de2551ece -> trunk/87329491c82a5f8c1cc4ec11d8f55a5de2551ece 2025-12-04T09:20:54.5843329Z * [new tag] trunk/892640e25aeefa8007c5af837214b4502b6b62a6 -> trunk/892640e25aeefa8007c5af837214b4502b6b62a6 2025-12-04T09:20:54.5843885Z * [new tag] trunk/89e3bbcb5b5321dc8b9520b4d5a8ee60cea1d0b4 -> trunk/89e3bbcb5b5321dc8b9520b4d5a8ee60cea1d0b4 2025-12-04T09:20:54.5844457Z * [new tag] trunk/8c73bbbb02159223c0c97d268a0a74cb78158a1c -> trunk/8c73bbbb02159223c0c97d268a0a74cb78158a1c 2025-12-04T09:20:54.5845107Z * [new tag] trunk/8d56e98c8db988a22cb2dfaeefb30bc7d2a3cc43 -> trunk/8d56e98c8db988a22cb2dfaeefb30bc7d2a3cc43 2025-12-04T09:20:54.5845687Z * [new tag] trunk/8d9dd9603e5ee26c01007f0cd4f018e584840922 -> trunk/8d9dd9603e5ee26c01007f0cd4f018e584840922 2025-12-04T09:20:54.5846258Z * [new tag] trunk/8ef0c0b02b062d75e7c9be2594914a3e784d23ca -> trunk/8ef0c0b02b062d75e7c9be2594914a3e784d23ca 2025-12-04T09:20:54.5846839Z * [new tag] trunk/90b27e7e8352cde97d32ddad24740ef819633f38 -> trunk/90b27e7e8352cde97d32ddad24740ef819633f38 2025-12-04T09:20:54.5847411Z * [new tag] trunk/90f0139e64b2951815d524b6a373bed20c4fbf90 -> trunk/90f0139e64b2951815d524b6a373bed20c4fbf90 2025-12-04T09:20:54.5847978Z * [new tag] trunk/93d0d6838c56af59b0dba794e6aa08f0c1c7799c -> trunk/93d0d6838c56af59b0dba794e6aa08f0c1c7799c 2025-12-04T09:20:54.5848558Z * [new tag] trunk/94ca8d5f1e81fea3ae488650a0fb6795049a9f87 -> trunk/94ca8d5f1e81fea3ae488650a0fb6795049a9f87 2025-12-04T09:20:54.5849137Z * [new tag] trunk/9844fbeadd5cebdf1281d6fbf79164139c352693 -> trunk/9844fbeadd5cebdf1281d6fbf79164139c352693 2025-12-04T09:20:54.5849721Z * [new tag] trunk/99024dec888ec1e50b546822a32b6fb2f35e5eaa -> trunk/99024dec888ec1e50b546822a32b6fb2f35e5eaa 2025-12-04T09:20:54.5850279Z * [new tag] trunk/9a296e640fc88aa44d275b48cd9cc30c573b169d -> trunk/9a296e640fc88aa44d275b48cd9cc30c573b169d 2025-12-04T09:20:54.5850865Z * [new tag] trunk/9b3e34d8589b29f7b4e7fab6f78711b7ca6e4639 -> trunk/9b3e34d8589b29f7b4e7fab6f78711b7ca6e4639 2025-12-04T09:20:54.5851423Z * [new tag] trunk/9cd055e547e9b67a5f9827f8999c38d7eda1bcb8 -> trunk/9cd055e547e9b67a5f9827f8999c38d7eda1bcb8 2025-12-04T09:20:54.5851999Z * [new tag] trunk/9f0df5686cb4ada94f94620acba2e3c3f363b11d -> trunk/9f0df5686cb4ada94f94620acba2e3c3f363b11d 2025-12-04T09:20:54.5852583Z * [new tag] trunk/9f7fceb887d0cfa0326a59b887821c63ff11340a -> trunk/9f7fceb887d0cfa0326a59b887821c63ff11340a 2025-12-04T09:20:54.5853147Z * [new tag] trunk/9f8ef8855d3078d70f7b782540ff2aaf158d6742 -> trunk/9f8ef8855d3078d70f7b782540ff2aaf158d6742 2025-12-04T09:20:54.5853745Z * [new tag] trunk/9fb52efc797b47a1f425a03aa5e47b866d8b1098 -> trunk/9fb52efc797b47a1f425a03aa5e47b866d8b1098 2025-12-04T09:20:54.5854296Z * [new tag] trunk/9ff4a2ebc5762d46c73e46b1b523d7ff349fedfa -> trunk/9ff4a2ebc5762d46c73e46b1b523d7ff349fedfa 2025-12-04T09:20:54.5854874Z * [new tag] trunk/a0f3937b94422354538ebbd47202d5b0e8a3fd0d -> trunk/a0f3937b94422354538ebbd47202d5b0e8a3fd0d 2025-12-04T09:20:54.5855409Z * [new tag] trunk/a15066c28b3145e6edbfc88359d0411d14cfc70c -> trunk/a15066c28b3145e6edbfc88359d0411d14cfc70c 2025-12-04T09:20:54.5855947Z * [new tag] trunk/a20f775e82564d2a9979221ed7f3b8d7cf54ce90 -> trunk/a20f775e82564d2a9979221ed7f3b8d7cf54ce90 2025-12-04T09:20:54.5856478Z * [new tag] trunk/a2973fb00ec002dd4b6bbf07385f066efb259b8c -> trunk/a2973fb00ec002dd4b6bbf07385f066efb259b8c 2025-12-04T09:20:54.5857030Z * [new tag] trunk/a7dc6dab9ad911259d4801c502907e531594db45 -> trunk/a7dc6dab9ad911259d4801c502907e531594db45 2025-12-04T09:20:54.5857565Z * [new tag] trunk/a951a9cee65c01660bbc6e6fded90ecb10fa6109 -> trunk/a951a9cee65c01660bbc6e6fded90ecb10fa6109 2025-12-04T09:20:54.5858112Z * [new tag] trunk/abfa1a6d65c7c159e35c72c25979b9da4971689e -> trunk/abfa1a6d65c7c159e35c72c25979b9da4971689e 2025-12-04T09:20:54.5858647Z * [new tag] trunk/ae3a2395bf66151078e2d201716f7d63ce1c6f3e -> trunk/ae3a2395bf66151078e2d201716f7d63ce1c6f3e 2025-12-04T09:20:54.5859185Z * [new tag] trunk/afdff7f0325080dedac44d080cb5a3b0e65e6c5e -> trunk/afdff7f0325080dedac44d080cb5a3b0e65e6c5e 2025-12-04T09:20:54.5859728Z * [new tag] trunk/b1aed4e7a72c03a38f44543aaea0dae2e9b76d48 -> trunk/b1aed4e7a72c03a38f44543aaea0dae2e9b76d48 2025-12-04T09:20:54.5860275Z * [new tag] trunk/b1decff555cd50e2123c8c6e25cc0d447c411f62 -> trunk/b1decff555cd50e2123c8c6e25cc0d447c411f62 2025-12-04T09:20:54.5860830Z * [new tag] trunk/b2b6b034c9fd08672c40e63ef243556ad4c49bd2 -> trunk/b2b6b034c9fd08672c40e63ef243556ad4c49bd2 2025-12-04T09:20:54.5861356Z * [new tag] trunk/b39813b4a04931682b0491adba2138d01d716d99 -> trunk/b39813b4a04931682b0491adba2138d01d716d99 2025-12-04T09:20:54.5861893Z * [new tag] trunk/b3a7edb2311367974cc7cd764cfb11a5d6758b24 -> trunk/b3a7edb2311367974cc7cd764cfb11a5d6758b24 2025-12-04T09:20:54.5862430Z * [new tag] trunk/b4cc1329c86acaef6d42c1fac7169b8d870ab0d7 -> trunk/b4cc1329c86acaef6d42c1fac7169b8d870ab0d7 2025-12-04T09:20:54.5862971Z * [new tag] trunk/b555c39217f765759954a4f9f9bd1e9b87bed11a -> trunk/b555c39217f765759954a4f9f9bd1e9b87bed11a 2025-12-04T09:20:54.5863511Z * [new tag] trunk/b6b6c80379388b7f9932c3e6a0f9907bf430e417 -> trunk/b6b6c80379388b7f9932c3e6a0f9907bf430e417 2025-12-04T09:20:54.5864055Z * [new tag] trunk/b6b6d912df0b6f4082f8e50b18bd1de1dd7325f4 -> trunk/b6b6d912df0b6f4082f8e50b18bd1de1dd7325f4 2025-12-04T09:20:54.5864604Z * [new tag] trunk/b7d60685f8cbc939b68a20871e90db67e729329b -> trunk/b7d60685f8cbc939b68a20871e90db67e729329b 2025-12-04T09:20:54.5865150Z * [new tag] trunk/b7f6b9a4fc6259f7af068f31868b3119bb1bac3e -> trunk/b7f6b9a4fc6259f7af068f31868b3119bb1bac3e 2025-12-04T09:20:54.5865701Z * [new tag] trunk/b8c4ba3593761e7b2a3ebd86f040fb07b47c02cf -> trunk/b8c4ba3593761e7b2a3ebd86f040fb07b47c02cf 2025-12-04T09:20:54.5866251Z * [new tag] trunk/b9c8f3a4884befb965ff42620ce44a71b04887f5 -> trunk/b9c8f3a4884befb965ff42620ce44a71b04887f5 2025-12-04T09:20:54.5866778Z * [new tag] trunk/ba1412546f3082c0958c077acc2025e4dbc33f1f -> trunk/ba1412546f3082c0958c077acc2025e4dbc33f1f 2025-12-04T09:20:54.5867320Z * [new tag] trunk/bac403c0b38c63bdbcc0c31f1c2b0bc0260f610f -> trunk/bac403c0b38c63bdbcc0c31f1c2b0bc0260f610f 2025-12-04T09:20:54.5867910Z * [new tag] trunk/bb3034198b459401fabeab254e1b99f0115046e2 -> trunk/bb3034198b459401fabeab254e1b99f0115046e2 2025-12-04T09:20:54.5868446Z * [new tag] trunk/bc39b2b3bc7a6e19a42e62bd576974035086fe55 -> trunk/bc39b2b3bc7a6e19a42e62bd576974035086fe55 2025-12-04T09:20:54.5869031Z * [new tag] trunk/bc43d5b297f207a11d83d77ddf0152bdaabe15a8 -> trunk/bc43d5b297f207a11d83d77ddf0152bdaabe15a8 2025-12-04T09:20:54.5869587Z * [new tag] trunk/bc6a4863c7246a6493d16d4ea6eee71ec07c6a09 -> trunk/bc6a4863c7246a6493d16d4ea6eee71ec07c6a09 2025-12-04T09:20:54.5870146Z * [new tag] trunk/bea4912944defdbcb8b061800caab6cbbbd01df5 -> trunk/bea4912944defdbcb8b061800caab6cbbbd01df5 2025-12-04T09:20:54.5870699Z * [new tag] trunk/c04e2c656f48d82d1521b867bbbf03967b9b7564 -> trunk/c04e2c656f48d82d1521b867bbbf03967b9b7564 2025-12-04T09:20:54.5871223Z * [new tag] trunk/c0660bcee27e7d7731634e274576a7081882bede -> trunk/c0660bcee27e7d7731634e274576a7081882bede 2025-12-04T09:20:54.5871770Z * [new tag] trunk/c178ed43d3d99cbefe84fbfb21d6f282b20d62ac -> trunk/c178ed43d3d99cbefe84fbfb21d6f282b20d62ac 2025-12-04T09:20:54.5872322Z * [new tag] trunk/c55b1e8f61d041ee436d697449eb028931d574fb -> trunk/c55b1e8f61d041ee436d697449eb028931d574fb 2025-12-04T09:20:54.5872872Z * [new tag] trunk/c6ae7579fe12fe75f1a8f7043a494c90567273f1 -> trunk/c6ae7579fe12fe75f1a8f7043a494c90567273f1 2025-12-04T09:20:54.5873420Z * [new tag] trunk/c8210e7d94bad5ae21ac389fa4ba8a463c76c4d0 -> trunk/c8210e7d94bad5ae21ac389fa4ba8a463c76c4d0 2025-12-04T09:20:54.5874024Z * [new tag] trunk/cc0853af42122f8185321f542616f4474e717f09 -> trunk/cc0853af42122f8185321f542616f4474e717f09 2025-12-04T09:20:54.5874577Z * [new tag] trunk/cddec6562eabfa390d014fa3741a5659cf9c94c9 -> trunk/cddec6562eabfa390d014fa3741a5659cf9c94c9 2025-12-04T09:20:54.5875158Z * [new tag] trunk/ce5e7e3bf1f4b69a4f4f93d288ba75b906df492a -> trunk/ce5e7e3bf1f4b69a4f4f93d288ba75b906df492a 2025-12-04T09:20:54.5875716Z * [new tag] trunk/d038b0130ec7c20ebcac219301292fd8e98a1ace -> trunk/d038b0130ec7c20ebcac219301292fd8e98a1ace 2025-12-04T09:20:54.5876273Z * [new tag] trunk/d16447dacaf2420ea175f0c275c75da951f57d39 -> trunk/d16447dacaf2420ea175f0c275c75da951f57d39 2025-12-04T09:20:54.5876822Z * [new tag] trunk/d19f1e8cab6810bb2e99141f9976665954c67a50 -> trunk/d19f1e8cab6810bb2e99141f9976665954c67a50 2025-12-04T09:20:54.5877389Z * [new tag] trunk/d1c9f03b2a5af4104721712f8cdffe9b4f340c01 -> trunk/d1c9f03b2a5af4104721712f8cdffe9b4f340c01 2025-12-04T09:20:54.5877941Z * [new tag] trunk/d40f4950f2b7f7aa380a22fe0f6166e71680fbcf -> trunk/d40f4950f2b7f7aa380a22fe0f6166e71680fbcf 2025-12-04T09:20:54.5878485Z * [new tag] trunk/d5038950bacfe36bbf24a47a455fe76901deb8e8 -> trunk/d5038950bacfe36bbf24a47a455fe76901deb8e8 2025-12-04T09:20:54.5879041Z * [new tag] trunk/d54ff42903c2ae0533931ff11d23b35f875bdb3d -> trunk/d54ff42903c2ae0533931ff11d23b35f875bdb3d 2025-12-04T09:20:54.5879599Z * [new tag] trunk/d76697633a2d2b9cced1ae21161849b33bfe7e47 -> trunk/d76697633a2d2b9cced1ae21161849b33bfe7e47 2025-12-04T09:20:54.5880162Z * [new tag] trunk/d78f52b199c547106d4cd9d2856dd0805c118bf1 -> trunk/d78f52b199c547106d4cd9d2856dd0805c118bf1 2025-12-04T09:20:54.5880696Z * [new tag] trunk/d8fd5c6eed28e5004150691d048a3f6785e19a8e -> trunk/d8fd5c6eed28e5004150691d048a3f6785e19a8e 2025-12-04T09:20:54.5881261Z * [new tag] trunk/d900f5e86745dec76713f4b0ef07005ef36b2f5a -> trunk/d900f5e86745dec76713f4b0ef07005ef36b2f5a 2025-12-04T09:20:54.5881918Z * [new tag] trunk/d973dc6b87d763859fe1c5bd1287e3b6b1c49d1b -> trunk/d973dc6b87d763859fe1c5bd1287e3b6b1c49d1b 2025-12-04T09:20:54.5882559Z * [new tag] trunk/d998c03304cb6ede76e1ed535b4ddeb6c2bf40ec -> trunk/d998c03304cb6ede76e1ed535b4ddeb6c2bf40ec 2025-12-04T09:20:54.5883117Z * [new tag] trunk/d9cb8a70833101dbbe16b99520cfbdd70d0a87bf -> trunk/d9cb8a70833101dbbe16b99520cfbdd70d0a87bf 2025-12-04T09:20:54.5883713Z * [new tag] trunk/d9d5e91b43f70eb8637af55db6856d49be391ffd -> trunk/d9d5e91b43f70eb8637af55db6856d49be391ffd 2025-12-04T09:20:54.5884270Z * [new tag] trunk/dd18a75336a4fbd7497955cc5665904724fce889 -> trunk/dd18a75336a4fbd7497955cc5665904724fce889 2025-12-04T09:20:54.5884822Z * [new tag] trunk/ded9bcd61a059bf723e6e84689552962b480ea77 -> trunk/ded9bcd61a059bf723e6e84689552962b480ea77 2025-12-04T09:20:54.5885380Z * [new tag] trunk/dfbd3714d15c37a7b83b322a6b60f997fc00f50c -> trunk/dfbd3714d15c37a7b83b322a6b60f997fc00f50c 2025-12-04T09:20:54.5885932Z * [new tag] trunk/e115f9f4e4b039f8e9a642aaa2bd8254a920541b -> trunk/e115f9f4e4b039f8e9a642aaa2bd8254a920541b 2025-12-04T09:20:54.5886491Z * [new tag] trunk/e3f24fd73ad74c6e7176687986436956c7c18235 -> trunk/e3f24fd73ad74c6e7176687986436956c7c18235 2025-12-04T09:20:54.5887040Z * [new tag] trunk/e7d24d3ff93d1503ba63860b7057438ad93f918e -> trunk/e7d24d3ff93d1503ba63860b7057438ad93f918e 2025-12-04T09:20:54.5887591Z * [new tag] trunk/ea7035f462a0d2830865ee86c832bd101e1427fc -> trunk/ea7035f462a0d2830865ee86c832bd101e1427fc 2025-12-04T09:20:54.5888135Z * [new tag] trunk/eabb7ad2128580ef674446027b95bcf4e21e8df3 -> trunk/eabb7ad2128580ef674446027b95bcf4e21e8df3 2025-12-04T09:20:54.5888693Z * [new tag] trunk/eb5c63652a33da42e7018c23df5f20a3eb4c6ccf -> trunk/eb5c63652a33da42e7018c23df5f20a3eb4c6ccf 2025-12-04T09:20:54.5889251Z * [new tag] trunk/ec2c71f5c85021b8938cdafadce24c15a36fd93e -> trunk/ec2c71f5c85021b8938cdafadce24c15a36fd93e 2025-12-04T09:20:54.5889818Z * [new tag] trunk/ecbcc3f6bf327856b435b259ac63cc2f328c4b4e -> trunk/ecbcc3f6bf327856b435b259ac63cc2f328c4b4e 2025-12-04T09:20:54.5890406Z * [new tag] trunk/ee87bbe876c42575e961b32a0827d76bc9782ca2 -> trunk/ee87bbe876c42575e961b32a0827d76bc9782ca2 2025-12-04T09:20:54.5890973Z * [new tag] trunk/ef019d1d431c4c5a95b594cb90d40a50cd00f5e4 -> trunk/ef019d1d431c4c5a95b594cb90d40a50cd00f5e4 2025-12-04T09:20:54.5891532Z * [new tag] trunk/ef8ecc13830a86c4b231f1aad9aba7851db61b53 -> trunk/ef8ecc13830a86c4b231f1aad9aba7851db61b53 2025-12-04T09:20:54.5892079Z * [new tag] trunk/f1076f5510920044912247b1abb8760cb820f598 -> trunk/f1076f5510920044912247b1abb8760cb820f598 2025-12-04T09:20:54.5892625Z * [new tag] trunk/f2d6a75a00a1d648ca9a0abc6a33e14c3dea6c40 -> trunk/f2d6a75a00a1d648ca9a0abc6a33e14c3dea6c40 2025-12-04T09:20:54.5893186Z * [new tag] trunk/f47dd0ddef1359e5b43e4b962412f67b30ecde56 -> trunk/f47dd0ddef1359e5b43e4b962412f67b30ecde56 2025-12-04T09:20:54.5893752Z * [new tag] trunk/f49d32dfa4730dcfb1b60eeeb369b5889da983c8 -> trunk/f49d32dfa4730dcfb1b60eeeb369b5889da983c8 2025-12-04T09:20:54.5894301Z * [new tag] trunk/f4dedf78fc30fd4b93975787ca6074ee89db9467 -> trunk/f4dedf78fc30fd4b93975787ca6074ee89db9467 2025-12-04T09:20:54.5894846Z * [new tag] trunk/f7c0d03819ebed05c4038f095d66d1b8c54aca17 -> trunk/f7c0d03819ebed05c4038f095d66d1b8c54aca17 2025-12-04T09:20:54.5895375Z * [new tag] trunk/f7e1bd80a063e17453c361837ba6ea2570920a73 -> trunk/f7e1bd80a063e17453c361837ba6ea2570920a73 2025-12-04T09:20:54.5895916Z * [new tag] trunk/f9bd6c53624c7c0ea3772de78498326e84c2f0e7 -> trunk/f9bd6c53624c7c0ea3772de78498326e84c2f0e7 2025-12-04T09:20:54.5896470Z * [new tag] trunk/fb5be221a46b51bfc9509013b0d85bc5a9d4f15b -> trunk/fb5be221a46b51bfc9509013b0d85bc5a9d4f15b 2025-12-04T09:20:54.5897011Z * [new tag] trunk/fdf863d5e1de3b2688c9511e96876e34581dbfd7 -> trunk/fdf863d5e1de3b2688c9511e96876e34581dbfd7 2025-12-04T09:20:54.5897599Z * [new tag] trunk/fe0e65adfc0e7ca6e5f57e6ea8b16bd5cc967307 -> trunk/fe0e65adfc0e7ca6e5f57e6ea8b16bd5cc967307 2025-12-04T09:20:54.5898148Z * [new tag] trunk/fec710bf89173f5355468a7ce1afe9157c3d9009 -> trunk/fec710bf89173f5355468a7ce1afe9157c3d9009 2025-12-04T09:20:54.5898744Z * [new tag] trunk/ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 -> trunk/ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:20:54.5899197Z * [new tag] v0.1.1 -> v0.1.1 2025-12-04T09:20:54.5899482Z * [new tag] v0.1.10 -> v0.1.10 2025-12-04T09:20:54.5899755Z * [new tag] v0.1.11 -> v0.1.11 2025-12-04T09:20:54.5900012Z * [new tag] v0.1.12 -> v0.1.12 2025-12-04T09:20:54.5900272Z * [new tag] v0.1.2 -> v0.1.2 2025-12-04T09:20:54.5900528Z * [new tag] v0.1.3 -> v0.1.3 2025-12-04T09:20:54.5900788Z * [new tag] v0.1.4 -> v0.1.4 2025-12-04T09:20:54.5901043Z * [new tag] v0.1.5 -> v0.1.5 2025-12-04T09:20:54.5901287Z * [new tag] v0.1.6 -> v0.1.6 2025-12-04T09:20:54.5901543Z * [new tag] v0.1.7 -> v0.1.7 2025-12-04T09:20:54.5901795Z * [new tag] v0.1.8 -> v0.1.8 2025-12-04T09:20:54.5902041Z * [new tag] v0.1.9 -> v0.1.9 2025-12-04T09:20:54.5902285Z * [new tag] v0.2.0 -> v0.2.0 2025-12-04T09:20:54.5902535Z * [new tag] v0.3.0 -> v0.3.0 2025-12-04T09:20:54.5902783Z * [new tag] v0.3.1 -> v0.3.1 2025-12-04T09:20:54.5903029Z * [new tag] v0.4.0 -> v0.4.0 2025-12-04T09:20:54.5903284Z * [new tag] v0.4.1 -> v0.4.1 2025-12-04T09:20:54.5903537Z * [new tag] v1.0.0 -> v1.0.0 2025-12-04T09:20:54.5903807Z * [new tag] v1.0.0a0 -> v1.0.0a0 2025-12-04T09:20:54.5904061Z * [new tag] v1.0.1 -> v1.0.1 2025-12-04T09:20:54.5904575Z * [new tag] v1.0rc0 -> v1.0rc0 2025-12-04T09:20:54.5904978Z * [new tag] v1.0rc1 -> v1.0rc1 2025-12-04T09:20:54.5905352Z * [new tag] v1.1.0 -> v1.1.0 2025-12-04T09:20:54.5905739Z * [new tag] v1.1.0a0 -> v1.1.0a0 2025-12-04T09:20:54.5906644Z * [new tag] v1.10.0 -> v1.10.0 2025-12-04T09:20:54.5906992Z * [new tag] v1.10.0-rc1 -> v1.10.0-rc1 2025-12-04T09:20:54.5907301Z * [new tag] v1.10.0-rc2 -> v1.10.0-rc2 2025-12-04T09:20:54.5907574Z * [new tag] v1.10.0-rc3 -> v1.10.0-rc3 2025-12-04T09:20:54.5907856Z * [new tag] v1.10.1 -> v1.10.1 2025-12-04T09:20:54.5908131Z * [new tag] v1.10.1-rc1 -> v1.10.1-rc1 2025-12-04T09:20:54.5909795Z * [new tag] v1.10.2 -> v1.10.2 2025-12-04T09:20:54.5913427Z * [new tag] v1.10.2-rc1 -> v1.10.2-rc1 2025-12-04T09:20:54.5918424Z * [new tag] v1.11.0 -> v1.11.0 2025-12-04T09:20:54.5920583Z * [new tag] v1.11.0-rc1 -> v1.11.0-rc1 2025-12-04T09:20:54.5920894Z * [new tag] v1.11.0-rc2 -> v1.11.0-rc2 2025-12-04T09:20:54.5921186Z * [new tag] v1.11.0-rc3 -> v1.11.0-rc3 2025-12-04T09:20:54.5921473Z * [new tag] v1.11.0-rc4 -> v1.11.0-rc4 2025-12-04T09:20:54.5922131Z * [new tag] v1.11.0-rc5 -> v1.11.0-rc5 2025-12-04T09:20:54.5922416Z * [new tag] v1.11.0-rc6 -> v1.11.0-rc6 2025-12-04T09:20:54.5922707Z * [new tag] v1.11.0-rc7 -> v1.11.0-rc7 2025-12-04T09:20:54.5923146Z * [new tag] v1.12.0 -> v1.12.0 2025-12-04T09:20:54.5923440Z * [new tag] v1.12.0-rc1 -> v1.12.0-rc1 2025-12-04T09:20:54.5923707Z * [new tag] v1.12.0-rc2 -> v1.12.0-rc2 2025-12-04T09:20:54.5923992Z * [new tag] v1.12.0-rc3 -> v1.12.0-rc3 2025-12-04T09:20:54.5924271Z * [new tag] v1.12.0-rc4 -> v1.12.0-rc4 2025-12-04T09:20:54.5924589Z * [new tag] v1.12.0-rc5 -> v1.12.0-rc5 2025-12-04T09:20:54.5924873Z * [new tag] v1.12.0-rc6 -> v1.12.0-rc6 2025-12-04T09:20:54.5925160Z * [new tag] v1.12.0-rc7 -> v1.12.0-rc7 2025-12-04T09:20:54.5925433Z * [new tag] v1.12.0-rc8 -> v1.12.0-rc8 2025-12-04T09:20:54.5925755Z * [new tag] v1.12.1 -> v1.12.1 2025-12-04T09:20:54.5926028Z * [new tag] v1.12.1-rc1 -> v1.12.1-rc1 2025-12-04T09:20:54.5926307Z * [new tag] v1.12.1-rc2 -> v1.12.1-rc2 2025-12-04T09:20:54.5926571Z * [new tag] v1.12.1-rc3 -> v1.12.1-rc3 2025-12-04T09:20:54.5926842Z * [new tag] v1.12.1-rc4 -> v1.12.1-rc4 2025-12-04T09:20:54.5927111Z * [new tag] v1.12.1-rc5 -> v1.12.1-rc5 2025-12-04T09:20:54.5927380Z * [new tag] v1.13.0 -> v1.13.0 2025-12-04T09:20:54.5927639Z * [new tag] v1.13.0-rc1 -> v1.13.0-rc1 2025-12-04T09:20:54.5927912Z * [new tag] v1.13.0-rc2 -> v1.13.0-rc2 2025-12-04T09:20:54.5928190Z * [new tag] v1.13.0-rc3 -> v1.13.0-rc3 2025-12-04T09:20:54.5928454Z * [new tag] v1.13.0-rc4 -> v1.13.0-rc4 2025-12-04T09:20:54.5933064Z * [new tag] v1.13.0-rc5 -> v1.13.0-rc5 2025-12-04T09:20:54.5939345Z * [new tag] v1.13.0-rc6 -> v1.13.0-rc6 2025-12-04T09:20:54.5944338Z * [new tag] v1.13.1 -> v1.13.1 2025-12-04T09:20:54.5946285Z * [new tag] v1.13.1-rc1 -> v1.13.1-rc1 2025-12-04T09:20:54.5946652Z * [new tag] v1.2.0 -> v1.2.0 2025-12-04T09:20:54.5947020Z * [new tag] v1.2.0a0 -> v1.2.0a0 2025-12-04T09:20:54.5947310Z * [new tag] v1.3.0 -> v1.3.0 2025-12-04T09:20:54.5947602Z * [new tag] v1.3.0a0 -> v1.3.0a0 2025-12-04T09:20:54.5947884Z * [new tag] v1.3.1 -> v1.3.1 2025-12-04T09:20:54.5948160Z * [new tag] v1.4.0 -> v1.4.0 2025-12-04T09:20:54.5948446Z * [new tag] v1.4.0a0 -> v1.4.0a0 2025-12-04T09:20:54.5948717Z * [new tag] v1.4.1 -> v1.4.1 2025-12-04T09:20:54.5948976Z * [new tag] v1.5.0 -> v1.5.0 2025-12-04T09:20:54.5949256Z * [new tag] v1.5.0-rc1 -> v1.5.0-rc1 2025-12-04T09:20:54.5949534Z * [new tag] v1.5.0-rc2 -> v1.5.0-rc2 2025-12-04T09:20:54.5949810Z * [new tag] v1.5.0-rc3 -> v1.5.0-rc3 2025-12-04T09:20:54.5950086Z * [new tag] v1.5.0-rc4 -> v1.5.0-rc4 2025-12-04T09:20:54.5950369Z * [new tag] v1.5.0-rc5 -> v1.5.0-rc5 2025-12-04T09:20:54.5950896Z * [new tag] v1.5.1 -> v1.5.1 2025-12-04T09:20:54.5951171Z * [new tag] v1.5.1-rc1 -> v1.5.1-rc1 2025-12-04T09:20:54.5951452Z * [new tag] v1.6.0 -> v1.6.0 2025-12-04T09:20:54.5951716Z * [new tag] v1.6.0-rc1 -> v1.6.0-rc1 2025-12-04T09:20:54.5952065Z * [new tag] v1.6.0-rc2 -> v1.6.0-rc2 2025-12-04T09:20:54.5952340Z * [new tag] v1.6.0-rc3 -> v1.6.0-rc3 2025-12-04T09:20:54.5952663Z * [new tag] v1.6.0-rc4 -> v1.6.0-rc4 2025-12-04T09:20:54.5952931Z * [new tag] v1.6.0-rc5 -> v1.6.0-rc5 2025-12-04T09:20:54.5953207Z * [new tag] v1.6.0-rc6 -> v1.6.0-rc6 2025-12-04T09:20:54.5953484Z * [new tag] v1.6.0-rc7 -> v1.6.0-rc7 2025-12-04T09:20:54.5953761Z * [new tag] v1.7.0 -> v1.7.0 2025-12-04T09:20:54.5954024Z * [new tag] v1.7.0-rc1 -> v1.7.0-rc1 2025-12-04T09:20:54.5954289Z * [new tag] v1.7.0-rc2 -> v1.7.0-rc2 2025-12-04T09:20:54.5954550Z * [new tag] v1.7.0-rc3 -> v1.7.0-rc3 2025-12-04T09:20:54.5954810Z * [new tag] v1.7.0-rc4 -> v1.7.0-rc4 2025-12-04T09:20:54.5955073Z * [new tag] v1.7.1 -> v1.7.1 2025-12-04T09:20:54.5955332Z * [new tag] v1.7.1-rc1 -> v1.7.1-rc1 2025-12-04T09:20:54.5955594Z * [new tag] v1.7.1-rc2 -> v1.7.1-rc2 2025-12-04T09:20:54.5955847Z * [new tag] v1.7.1-rc3 -> v1.7.1-rc3 2025-12-04T09:20:54.5956107Z * [new tag] v1.8.0 -> v1.8.0 2025-12-04T09:20:54.5956368Z * [new tag] v1.8.0-rc1 -> v1.8.0-rc1 2025-12-04T09:20:54.5956632Z * [new tag] v1.8.0-rc2 -> v1.8.0-rc2 2025-12-04T09:20:54.5956908Z * [new tag] v1.8.0-rc3 -> v1.8.0-rc3 2025-12-04T09:20:54.5957236Z * [new tag] v1.8.0-rc4 -> v1.8.0-rc4 2025-12-04T09:20:54.5957518Z * [new tag] v1.8.0-rc5 -> v1.8.0-rc5 2025-12-04T09:20:54.5957797Z * [new tag] v1.8.1 -> v1.8.1 2025-12-04T09:20:54.5958063Z * [new tag] v1.8.1-rc1 -> v1.8.1-rc1 2025-12-04T09:20:54.5958345Z * [new tag] v1.8.1-rc2 -> v1.8.1-rc2 2025-12-04T09:20:54.5958624Z * [new tag] v1.8.1-rc3 -> v1.8.1-rc3 2025-12-04T09:20:54.5958892Z * [new tag] v1.8.2 -> v1.8.2 2025-12-04T09:20:54.5959154Z * [new tag] v1.8.2-rc1 -> v1.8.2-rc1 2025-12-04T09:20:54.5959421Z * [new tag] v1.9.0 -> v1.9.0 2025-12-04T09:20:54.5959678Z * [new tag] v1.9.0-rc1 -> v1.9.0-rc1 2025-12-04T09:20:54.5959957Z * [new tag] v1.9.0-rc2 -> v1.9.0-rc2 2025-12-04T09:20:54.5960236Z * [new tag] v1.9.0-rc3 -> v1.9.0-rc3 2025-12-04T09:20:54.5960517Z * [new tag] v1.9.0-rc4 -> v1.9.0-rc4 2025-12-04T09:20:54.5960796Z * [new tag] v1.9.1 -> v1.9.1 2025-12-04T09:20:54.5961057Z * [new tag] v1.9.1-rc1 -> v1.9.1-rc1 2025-12-04T09:20:54.5961334Z * [new tag] v1.9.1-rc2 -> v1.9.1-rc2 2025-12-04T09:20:54.5961612Z * [new tag] v2.0.0 -> v2.0.0 2025-12-04T09:20:54.5962125Z * [new tag] v2.0.0-rc1 -> v2.0.0-rc1 2025-12-04T09:20:54.5962486Z * [new tag] v2.0.0-rc2 -> v2.0.0-rc2 2025-12-04T09:20:54.5962769Z * [new tag] v2.0.0-rc3 -> v2.0.0-rc3 2025-12-04T09:20:54.5963044Z * [new tag] v2.0.0-rc4 -> v2.0.0-rc4 2025-12-04T09:20:54.5963314Z * [new tag] v2.0.0-rc5 -> v2.0.0-rc5 2025-12-04T09:20:54.5963726Z * [new tag] v2.0.0-rc6 -> v2.0.0-rc6 2025-12-04T09:20:54.5964009Z * [new tag] v2.0.1 -> v2.0.1 2025-12-04T09:20:54.5964110Z * [new tag] v2.0.1-rc1 -> v2.0.1-rc1 2025-12-04T09:20:54.5964210Z * [new tag] v2.0.1-rc2 -> v2.0.1-rc2 2025-12-04T09:20:54.5964319Z * [new tag] v2.0.1-rc3 -> v2.0.1-rc3 2025-12-04T09:20:54.5964426Z * [new tag] v2.0.1-rc4 -> v2.0.1-rc4 2025-12-04T09:20:54.5966066Z * [new tag] v2.1.0 -> v2.1.0 2025-12-04T09:20:54.5968275Z * [new tag] v2.1.0-rc1 -> v2.1.0-rc1 2025-12-04T09:20:54.5968453Z * [new tag] v2.1.0-rc2 -> v2.1.0-rc2 2025-12-04T09:20:54.5968582Z * [new tag] v2.1.0-rc3 -> v2.1.0-rc3 2025-12-04T09:20:54.5968707Z * [new tag] v2.1.0-rc4 -> v2.1.0-rc4 2025-12-04T09:20:54.5968832Z * [new tag] v2.1.0-rc5 -> v2.1.0-rc5 2025-12-04T09:20:54.5969177Z * [new tag] v2.1.0-rc6 -> v2.1.0-rc6 2025-12-04T09:20:54.5969485Z * [new tag] v2.1.1 -> v2.1.1 2025-12-04T09:20:54.5970043Z * [new tag] v2.1.1-rc1 -> v2.1.1-rc1 2025-12-04T09:20:54.5971424Z * [new tag] v2.1.1-rc2 -> v2.1.1-rc2 2025-12-04T09:20:54.5971550Z * [new tag] v2.1.1-rc3 -> v2.1.1-rc3 2025-12-04T09:20:54.5971887Z * [new tag] v2.1.1-rc4 -> v2.1.1-rc4 2025-12-04T09:20:54.5974102Z * [new tag] v2.1.1-rc5 -> v2.1.1-rc5 2025-12-04T09:20:54.5974254Z * [new tag] v2.1.1-rc6 -> v2.1.1-rc6 2025-12-04T09:20:54.5974370Z * [new tag] v2.1.2 -> v2.1.2 2025-12-04T09:20:54.5974952Z * [new tag] v2.1.2-rc1 -> v2.1.2-rc1 2025-12-04T09:20:54.5975165Z * [new tag] v2.1.2-rc2 -> v2.1.2-rc2 2025-12-04T09:20:54.5975279Z * [new tag] v2.1.2-rc3 -> v2.1.2-rc3 2025-12-04T09:20:54.5976165Z * [new tag] v2.2.0 -> v2.2.0 2025-12-04T09:20:54.5976542Z * [new tag] v2.2.0-rc1 -> v2.2.0-rc1 2025-12-04T09:20:54.5979023Z * [new tag] v2.2.0-rc2 -> v2.2.0-rc2 2025-12-04T09:20:54.5979185Z * [new tag] v2.2.0-rc3 -> v2.2.0-rc3 2025-12-04T09:20:54.5979310Z * [new tag] v2.2.0-rc4 -> v2.2.0-rc4 2025-12-04T09:20:54.5979417Z * [new tag] v2.2.0-rc5 -> v2.2.0-rc5 2025-12-04T09:20:54.5984529Z * [new tag] v2.2.0-rc6 -> v2.2.0-rc6 2025-12-04T09:20:54.5984682Z * [new tag] v2.2.0-rc7 -> v2.2.0-rc7 2025-12-04T09:20:54.5984795Z * [new tag] v2.2.0-rc8 -> v2.2.0-rc8 2025-12-04T09:20:54.5984926Z * [new tag] v2.2.1 -> v2.2.1 2025-12-04T09:20:54.5985037Z * [new tag] v2.2.1-rc1 -> v2.2.1-rc1 2025-12-04T09:20:54.5985148Z * [new tag] v2.2.1-rc2 -> v2.2.1-rc2 2025-12-04T09:20:54.5985254Z * [new tag] v2.2.1-rc3 -> v2.2.1-rc3 2025-12-04T09:20:54.5985518Z * [new tag] v2.2.2 -> v2.2.2 2025-12-04T09:20:54.5985633Z * [new tag] v2.2.2-rc1 -> v2.2.2-rc1 2025-12-04T09:20:54.5985737Z * [new tag] v2.2.2-rc2 -> v2.2.2-rc2 2025-12-04T09:20:54.5986262Z * [new tag] v2.2.2-rc3 -> v2.2.2-rc3 2025-12-04T09:20:54.5986755Z * [new tag] v2.3.0 -> v2.3.0 2025-12-04T09:20:54.5986888Z * [new tag] v2.3.0-rc1 -> v2.3.0-rc1 2025-12-04T09:20:54.5987339Z * [new tag] v2.3.0-rc10 -> v2.3.0-rc10 2025-12-04T09:20:54.5988211Z * [new tag] v2.3.0-rc11 -> v2.3.0-rc11 2025-12-04T09:20:54.5988376Z * [new tag] v2.3.0-rc12 -> v2.3.0-rc12 2025-12-04T09:20:54.5990797Z * [new tag] v2.3.0-rc2 -> v2.3.0-rc2 2025-12-04T09:20:54.5991334Z * [new tag] v2.3.0-rc3 -> v2.3.0-rc3 2025-12-04T09:20:54.5991472Z * [new tag] v2.3.0-rc4 -> v2.3.0-rc4 2025-12-04T09:20:54.5991578Z * [new tag] v2.3.0-rc5 -> v2.3.0-rc5 2025-12-04T09:20:54.5991699Z * [new tag] v2.3.0-rc6 -> v2.3.0-rc6 2025-12-04T09:20:54.5992117Z * [new tag] v2.3.0-rc7 -> v2.3.0-rc7 2025-12-04T09:20:54.5995315Z * [new tag] v2.3.0-rc8 -> v2.3.0-rc8 2025-12-04T09:20:54.5999661Z * [new tag] v2.3.0-rc9 -> v2.3.0-rc9 2025-12-04T09:20:54.5999817Z * [new tag] v2.3.1 -> v2.3.1 2025-12-04T09:20:54.5999934Z * [new tag] v2.3.1-rc1 -> v2.3.1-rc1 2025-12-04T09:20:54.6000046Z * [new tag] v2.3.1-rc2 -> v2.3.1-rc2 2025-12-04T09:20:54.6000151Z * [new tag] v2.3.1-rc3 -> v2.3.1-rc3 2025-12-04T09:20:54.6000271Z * [new tag] v2.4.0 -> v2.4.0 2025-12-04T09:20:54.6000382Z * [new tag] v2.4.0-rc1 -> v2.4.0-rc1 2025-12-04T09:20:54.6000484Z * [new tag] v2.4.0-rc2 -> v2.4.0-rc2 2025-12-04T09:20:54.6000593Z * [new tag] v2.4.0-rc3 -> v2.4.0-rc3 2025-12-04T09:20:54.6000703Z * [new tag] v2.4.0-rc4 -> v2.4.0-rc4 2025-12-04T09:20:54.6000813Z * [new tag] v2.4.0-rc5 -> v2.4.0-rc5 2025-12-04T09:20:54.6000921Z * [new tag] v2.4.0-rc6 -> v2.4.0-rc6 2025-12-04T09:20:54.6001062Z * [new tag] v2.4.0-rc7 -> v2.4.0-rc7 2025-12-04T09:20:54.6001571Z * [new tag] v2.4.0-rc8 -> v2.4.0-rc8 2025-12-04T09:20:54.6007350Z * [new tag] v2.4.0-rc9 -> v2.4.0-rc9 2025-12-04T09:20:54.6009318Z * [new tag] v2.4.1 -> v2.4.1 2025-12-04T09:20:54.6009571Z * [new tag] v2.4.1-rc1 -> v2.4.1-rc1 2025-12-04T09:20:54.6015454Z * [new tag] v2.4.1-rc2 -> v2.4.1-rc2 2025-12-04T09:20:54.6020516Z * [new tag] v2.4.1-rc3 -> v2.4.1-rc3 2025-12-04T09:20:54.6026199Z * [new tag] v2.5.0 -> v2.5.0 2025-12-04T09:20:54.6026500Z * [new tag] v2.5.0-rc1 -> v2.5.0-rc1 2025-12-04T09:20:54.6026650Z * [new tag] v2.5.0-rc10 -> v2.5.0-rc10 2025-12-04T09:20:54.6026850Z * [new tag] v2.5.0-rc2 -> v2.5.0-rc2 2025-12-04T09:20:54.6026980Z * [new tag] v2.5.0-rc3 -> v2.5.0-rc3 2025-12-04T09:20:54.6027121Z * [new tag] v2.5.0-rc4 -> v2.5.0-rc4 2025-12-04T09:20:54.6027513Z * [new tag] v2.5.0-rc5 -> v2.5.0-rc5 2025-12-04T09:20:54.6027624Z * [new tag] v2.5.0-rc6 -> v2.5.0-rc6 2025-12-04T09:20:54.6028248Z * [new tag] v2.5.0-rc7 -> v2.5.0-rc7 2025-12-04T09:20:54.6028400Z * [new tag] v2.5.0-rc8 -> v2.5.0-rc8 2025-12-04T09:20:54.6028673Z * [new tag] v2.5.0-rc9 -> v2.5.0-rc9 2025-12-04T09:20:54.6028789Z * [new tag] v2.5.1 -> v2.5.1 2025-12-04T09:20:54.6028901Z * [new tag] v2.5.1-rc1 -> v2.5.1-rc1 2025-12-04T09:20:54.6029002Z * [new tag] v2.6.0 -> v2.6.0 2025-12-04T09:20:54.6029104Z * [new tag] v2.6.0-rc1 -> v2.6.0-rc1 2025-12-04T09:20:54.6029215Z * [new tag] v2.6.0-rc2 -> v2.6.0-rc2 2025-12-04T09:20:54.6029309Z * [new tag] v2.6.0-rc3 -> v2.6.0-rc3 2025-12-04T09:20:54.6029411Z * [new tag] v2.6.0-rc4 -> v2.6.0-rc4 2025-12-04T09:20:54.6029516Z * [new tag] v2.6.0-rc5 -> v2.6.0-rc5 2025-12-04T09:20:54.6029611Z * [new tag] v2.6.0-rc6 -> v2.6.0-rc6 2025-12-04T09:20:54.6029711Z * [new tag] v2.6.0-rc7 -> v2.6.0-rc7 2025-12-04T09:20:54.6029806Z * [new tag] v2.6.0-rc8 -> v2.6.0-rc8 2025-12-04T09:20:54.6029898Z * [new tag] v2.6.0-rc9 -> v2.6.0-rc9 2025-12-04T09:20:54.6029998Z * [new tag] v2.7.0 -> v2.7.0 2025-12-04T09:20:54.6030089Z * [new tag] v2.7.0-rc1 -> v2.7.0-rc1 2025-12-04T09:20:54.6030361Z * [new tag] v2.7.0-rc10 -> v2.7.0-rc10 2025-12-04T09:20:54.6030472Z * [new tag] v2.7.0-rc2 -> v2.7.0-rc2 2025-12-04T09:20:54.6030572Z * [new tag] v2.7.0-rc3 -> v2.7.0-rc3 2025-12-04T09:20:54.6030674Z * [new tag] v2.7.0-rc4 -> v2.7.0-rc4 2025-12-04T09:20:54.6030769Z * [new tag] v2.7.0-rc5 -> v2.7.0-rc5 2025-12-04T09:20:54.6030864Z * [new tag] v2.7.0-rc6 -> v2.7.0-rc6 2025-12-04T09:20:54.6030972Z * [new tag] v2.7.0-rc7 -> v2.7.0-rc7 2025-12-04T09:20:54.6031067Z * [new tag] v2.7.0-rc8 -> v2.7.0-rc8 2025-12-04T09:20:54.6031170Z * [new tag] v2.7.0-rc9 -> v2.7.0-rc9 2025-12-04T09:20:54.6031263Z * [new tag] v2.7.1 -> v2.7.1 2025-12-04T09:20:54.6031356Z * [new tag] v2.7.1-rc1 -> v2.7.1-rc1 2025-12-04T09:20:54.6031458Z * [new tag] v2.7.1-rc2 -> v2.7.1-rc2 2025-12-04T09:20:54.6031554Z * [new tag] v2.7.1-rc3 -> v2.7.1-rc3 2025-12-04T09:20:54.6031820Z * [new tag] v2.7.1-rc4 -> v2.7.1-rc4 2025-12-04T09:20:54.6031942Z * [new tag] v2.7.1-rc5 -> v2.7.1-rc5 2025-12-04T09:20:54.6032053Z * [new tag] v2.8.0 -> v2.8.0 2025-12-04T09:20:54.6032231Z * [new tag] v2.8.0-rc1 -> v2.8.0-rc1 2025-12-04T09:20:54.6032340Z * [new tag] v2.8.0-rc2 -> v2.8.0-rc2 2025-12-04T09:20:54.6032499Z * [new tag] v2.8.0-rc3 -> v2.8.0-rc3 2025-12-04T09:20:54.6032606Z * [new tag] v2.8.0-rc4 -> v2.8.0-rc4 2025-12-04T09:20:54.6038923Z * [new tag] v2.8.0-rc5 -> v2.8.0-rc5 2025-12-04T09:20:54.6039203Z * [new tag] v2.8.0-rc6 -> v2.8.0-rc6 2025-12-04T09:20:54.6039369Z * [new tag] v2.8.0-rc7 -> v2.8.0-rc7 2025-12-04T09:20:54.6039827Z * [new tag] v2.8.0-rc8 -> v2.8.0-rc8 2025-12-04T09:20:54.6039966Z * [new tag] v2.9.0 -> v2.9.0 2025-12-04T09:20:54.6040095Z * [new tag] v2.9.0-rc1 -> v2.9.0-rc1 2025-12-04T09:20:54.6040288Z * [new tag] v2.9.0-rc10 -> v2.9.0-rc10 2025-12-04T09:20:54.6040703Z * [new tag] v2.9.0-rc11 -> v2.9.0-rc11 2025-12-04T09:20:54.6042078Z * [new tag] v2.9.0-rc2 -> v2.9.0-rc2 2025-12-04T09:20:54.6047356Z * [new tag] v2.9.0-rc3 -> v2.9.0-rc3 2025-12-04T09:20:54.6053136Z * [new tag] v2.9.0-rc4 -> v2.9.0-rc4 2025-12-04T09:20:54.6058687Z * [new tag] v2.9.0-rc5 -> v2.9.0-rc5 2025-12-04T09:20:54.6063545Z * [new tag] v2.9.0-rc6 -> v2.9.0-rc6 2025-12-04T09:20:54.6068574Z * [new tag] v2.9.0-rc7 -> v2.9.0-rc7 2025-12-04T09:20:54.6072854Z * [new tag] v2.9.0-rc8 -> v2.9.0-rc8 2025-12-04T09:20:54.6077365Z * [new tag] v2.9.0-rc9 -> v2.9.0-rc9 2025-12-04T09:20:54.6079890Z * [new tag] v2.9.1 -> v2.9.1 2025-12-04T09:20:54.6080029Z * [new tag] v2.9.1-rc1 -> v2.9.1-rc1 2025-12-04T09:20:54.6080426Z * [new tag] v2.9.1-rc2 -> v2.9.1-rc2 2025-12-04T09:20:54.6080589Z * [new tag] viable/strict/1759343184 -> viable/strict/1759343184 2025-12-04T09:20:54.6080739Z * [new tag] viable/strict/1759346540 -> viable/strict/1759346540 2025-12-04T09:20:54.6080873Z * [new tag] viable/strict/1759348181 -> viable/strict/1759348181 2025-12-04T09:20:54.6081013Z * [new tag] viable/strict/1759350324 -> viable/strict/1759350324 2025-12-04T09:20:54.6081167Z * [new tag] viable/strict/1759351793 -> viable/strict/1759351793 2025-12-04T09:20:54.6081305Z * [new tag] viable/strict/1759353844 -> viable/strict/1759353844 2025-12-04T09:20:54.6081446Z * [new tag] viable/strict/1759355374 -> viable/strict/1759355374 2025-12-04T09:20:54.6081598Z * [new tag] viable/strict/1759357472 -> viable/strict/1759357472 2025-12-04T09:20:54.6081925Z * [new tag] viable/strict/1759361002 -> viable/strict/1759361002 2025-12-04T09:20:54.6082079Z * [new tag] viable/strict/1759362585 -> viable/strict/1759362585 2025-12-04T09:20:54.6082214Z * [new tag] viable/strict/1759365359 -> viable/strict/1759365359 2025-12-04T09:20:54.6082354Z * [new tag] viable/strict/1759370089 -> viable/strict/1759370089 2025-12-04T09:20:54.6082554Z * [new tag] viable/strict/1759377554 -> viable/strict/1759377554 2025-12-04T09:20:54.6082738Z * [new tag] viable/strict/1759379133 -> viable/strict/1759379133 2025-12-04T09:20:54.6082907Z * [new tag] viable/strict/1759389871 -> viable/strict/1759389871 2025-12-04T09:20:54.6083040Z * [new tag] viable/strict/1759393562 -> viable/strict/1759393562 2025-12-04T09:20:54.6083182Z * [new tag] viable/strict/1759395076 -> viable/strict/1759395076 2025-12-04T09:20:54.6083317Z * [new tag] viable/strict/1759398579 -> viable/strict/1759398579 2025-12-04T09:20:54.6083449Z * [new tag] viable/strict/1759404142 -> viable/strict/1759404142 2025-12-04T09:20:54.6083592Z * [new tag] viable/strict/1759405773 -> viable/strict/1759405773 2025-12-04T09:20:54.6083727Z * [new tag] viable/strict/1759408041 -> viable/strict/1759408041 2025-12-04T09:20:54.6083864Z * [new tag] viable/strict/1759411593 -> viable/strict/1759411593 2025-12-04T09:20:54.6084152Z * [new tag] viable/strict/1759427395 -> viable/strict/1759427395 2025-12-04T09:20:54.6084296Z * [new tag] viable/strict/1759434582 -> viable/strict/1759434582 2025-12-04T09:20:54.6084436Z * [new tag] viable/strict/1759436720 -> viable/strict/1759436720 2025-12-04T09:20:54.6084615Z * [new tag] viable/strict/1759440219 -> viable/strict/1759440219 2025-12-04T09:20:54.6084756Z * [new tag] viable/strict/1759441948 -> viable/strict/1759441948 2025-12-04T09:20:54.6084890Z * [new tag] viable/strict/1759443860 -> viable/strict/1759443860 2025-12-04T09:20:54.6085021Z * [new tag] viable/strict/1759445377 -> viable/strict/1759445377 2025-12-04T09:20:54.6085166Z * [new tag] viable/strict/1759447415 -> viable/strict/1759447415 2025-12-04T09:20:54.6085298Z * [new tag] viable/strict/1759451750 -> viable/strict/1759451750 2025-12-04T09:20:54.6085434Z * [new tag] viable/strict/1759453910 -> viable/strict/1759453910 2025-12-04T09:20:54.6085571Z * [new tag] viable/strict/1759456483 -> viable/strict/1759456483 2025-12-04T09:20:54.6085706Z * [new tag] viable/strict/1759459279 -> viable/strict/1759459279 2025-12-04T09:20:54.6085850Z * [new tag] viable/strict/1759460742 -> viable/strict/1759460742 2025-12-04T09:20:54.6085973Z * [new tag] viable/strict/1759462025 -> viable/strict/1759462025 2025-12-04T09:20:54.6086107Z * [new tag] viable/strict/1759469086 -> viable/strict/1759469086 2025-12-04T09:20:54.6086254Z * [new tag] viable/strict/1759470581 -> viable/strict/1759470581 2025-12-04T09:20:54.6086392Z * [new tag] viable/strict/1759472786 -> viable/strict/1759472786 2025-12-04T09:20:54.6086534Z * [new tag] viable/strict/1759476294 -> viable/strict/1759476294 2025-12-04T09:20:54.6086663Z * [new tag] viable/strict/1759479963 -> viable/strict/1759479963 2025-12-04T09:20:54.6086797Z * [new tag] viable/strict/1759492177 -> viable/strict/1759492177 2025-12-04T09:20:54.6086931Z * [new tag] viable/strict/1759519278 -> viable/strict/1759519278 2025-12-04T09:20:54.6087072Z * [new tag] viable/strict/1759524580 -> viable/strict/1759524580 2025-12-04T09:20:54.6087203Z * [new tag] viable/strict/1759528193 -> viable/strict/1759528193 2025-12-04T09:20:54.6087328Z * [new tag] viable/strict/1759533797 -> viable/strict/1759533797 2025-12-04T09:20:54.6087470Z * [new tag] viable/strict/1759542780 -> viable/strict/1759542780 2025-12-04T09:20:54.6087615Z * [new tag] viable/strict/1759549779 -> viable/strict/1759549779 2025-12-04T09:20:54.6087753Z * [new tag] viable/strict/1759555455 -> viable/strict/1759555455 2025-12-04T09:20:54.6087884Z * [new tag] viable/strict/1759559176 -> viable/strict/1759559176 2025-12-04T09:20:54.6088031Z * [new tag] viable/strict/1759560629 -> viable/strict/1759560629 2025-12-04T09:20:54.6088171Z * [new tag] viable/strict/1759569848 -> viable/strict/1759569848 2025-12-04T09:20:54.6088318Z * [new tag] viable/strict/1759571382 -> viable/strict/1759571382 2025-12-04T09:20:54.6088445Z * [new tag] viable/strict/1759573474 -> viable/strict/1759573474 2025-12-04T09:20:54.6088581Z * [new tag] viable/strict/1759618187 -> viable/strict/1759618187 2025-12-04T09:20:54.6088731Z * [new tag] viable/strict/1759626742 -> viable/strict/1759626742 2025-12-04T09:20:54.6088870Z * [new tag] viable/strict/1759632427 -> viable/strict/1759632427 2025-12-04T09:20:54.6089016Z * [new tag] viable/strict/1759634971 -> viable/strict/1759634971 2025-12-04T09:20:54.6089179Z * [new tag] viable/strict/1759661382 -> viable/strict/1759661382 2025-12-04T09:20:54.6089318Z * [new tag] viable/strict/1759663294 -> viable/strict/1759663294 2025-12-04T09:20:54.6089465Z * [new tag] viable/strict/1759708178 -> viable/strict/1759708178 2025-12-04T09:20:54.6089637Z * [new tag] viable/strict/1759715695 -> viable/strict/1759715695 2025-12-04T09:20:54.6089770Z * [new tag] viable/strict/1759728293 -> viable/strict/1759728293 2025-12-04T09:20:54.6089909Z * [new tag] viable/strict/1759735513 -> viable/strict/1759735513 2025-12-04T09:20:54.6090032Z * [new tag] viable/strict/1759739177 -> viable/strict/1759739177 2025-12-04T09:20:54.6090178Z * [new tag] viable/strict/1759758635 -> viable/strict/1759758635 2025-12-04T09:20:54.6090318Z * [new tag] viable/strict/1759765784 -> viable/strict/1759765784 2025-12-04T09:20:54.6090444Z * [new tag] viable/strict/1759767948 -> viable/strict/1759767948 2025-12-04T09:20:54.6090591Z * [new tag] viable/strict/1759771461 -> viable/strict/1759771461 2025-12-04T09:20:54.6090730Z * [new tag] viable/strict/1759776706 -> viable/strict/1759776706 2025-12-04T09:20:54.6096834Z * [new tag] viable/strict/1759782317 -> viable/strict/1759782317 2025-12-04T09:20:54.6102358Z * [new tag] viable/strict/1759783777 -> viable/strict/1759783777 2025-12-04T09:20:54.6104793Z * [new tag] viable/strict/1759785815 -> viable/strict/1759785815 2025-12-04T09:20:54.6105069Z * [new tag] viable/strict/1759789459 -> viable/strict/1759789459 2025-12-04T09:20:54.6111671Z * [new tag] viable/strict/1759790974 -> viable/strict/1759790974 2025-12-04T09:20:54.6115164Z * [new tag] viable/strict/1759794583 -> viable/strict/1759794583 2025-12-04T09:20:54.6115475Z * [new tag] viable/strict/1759797408 -> viable/strict/1759797408 2025-12-04T09:20:54.6115965Z * [new tag] viable/strict/1759799518 -> viable/strict/1759799518 2025-12-04T09:20:54.6116122Z * [new tag] viable/strict/1759804909 -> viable/strict/1759804909 2025-12-04T09:20:54.6116303Z * [new tag] viable/strict/1759807643 -> viable/strict/1759807643 2025-12-04T09:20:54.6116434Z * [new tag] viable/strict/1759809089 -> viable/strict/1759809089 2025-12-04T09:20:54.6116569Z * [new tag] viable/strict/1759811145 -> viable/strict/1759811145 2025-12-04T09:20:54.6116701Z * [new tag] viable/strict/1759812581 -> viable/strict/1759812581 2025-12-04T09:20:54.6116830Z * [new tag] viable/strict/1759814683 -> viable/strict/1759814683 2025-12-04T09:20:54.6116997Z * [new tag] viable/strict/1759821889 -> viable/strict/1759821889 2025-12-04T09:20:54.6117137Z * [new tag] viable/strict/1759823376 -> viable/strict/1759823376 2025-12-04T09:20:54.6117274Z * [new tag] viable/strict/1759827107 -> viable/strict/1759827107 2025-12-04T09:20:54.6117405Z * [new tag] viable/strict/1759830577 -> viable/strict/1759830577 2025-12-04T09:20:54.6117539Z * [new tag] viable/strict/1759832720 -> viable/strict/1759832720 2025-12-04T09:20:54.6117673Z * [new tag] viable/strict/1759842063 -> viable/strict/1759842063 2025-12-04T09:20:54.6117800Z * [new tag] viable/strict/1759847121 -> viable/strict/1759847121 2025-12-04T09:20:54.6117944Z * [new tag] viable/strict/1759850721 -> viable/strict/1759850721 2025-12-04T09:20:54.6118079Z * [new tag] viable/strict/1759857870 -> viable/strict/1759857870 2025-12-04T09:20:54.6118215Z * [new tag] viable/strict/1759863143 -> viable/strict/1759863143 2025-12-04T09:20:54.6118497Z * [new tag] viable/strict/1759875874 -> viable/strict/1759875874 2025-12-04T09:20:54.6118633Z * [new tag] viable/strict/1759877385 -> viable/strict/1759877385 2025-12-04T09:20:54.6118765Z * [new tag] viable/strict/1759883801 -> viable/strict/1759883801 2025-12-04T09:20:54.6118959Z * [new tag] viable/strict/1759885922 -> viable/strict/1759885922 2025-12-04T09:20:54.6119099Z * [new tag] viable/strict/1759888488 -> viable/strict/1759888488 2025-12-04T09:20:54.6119241Z * [new tag] viable/strict/1759895471 -> viable/strict/1759895471 2025-12-04T09:20:54.6119379Z * [new tag] viable/strict/1759904803 -> viable/strict/1759904803 2025-12-04T09:20:54.6119515Z * [new tag] viable/strict/1759908300 -> viable/strict/1759908300 2025-12-04T09:20:54.6119658Z * [new tag] viable/strict/1759915520 -> viable/strict/1759915520 2025-12-04T09:20:54.6119798Z * [new tag] viable/strict/1759916978 -> viable/strict/1759916978 2025-12-04T09:20:54.6119939Z * [new tag] viable/strict/1759930024 -> viable/strict/1759930024 2025-12-04T09:20:54.6120076Z * [new tag] viable/strict/1759948122 -> viable/strict/1759948122 2025-12-04T09:20:54.6120215Z * [new tag] viable/strict/1759952983 -> viable/strict/1759952983 2025-12-04T09:20:54.6120355Z * [new tag] viable/strict/1759955121 -> viable/strict/1759955121 2025-12-04T09:20:54.6120493Z * [new tag] viable/strict/1759962298 -> viable/strict/1759962298 2025-12-04T09:20:54.6120631Z * [new tag] viable/strict/1759965837 -> viable/strict/1759965837 2025-12-04T09:20:54.6120776Z * [new tag] viable/strict/1759970213 -> viable/strict/1759970213 2025-12-04T09:20:54.6120913Z * [new tag] viable/strict/1759974894 -> viable/strict/1759974894 2025-12-04T09:20:54.6121056Z * [new tag] viable/strict/1759977763 -> viable/strict/1759977763 2025-12-04T09:20:54.6121192Z * [new tag] viable/strict/1759979241 -> viable/strict/1759979241 2025-12-04T09:20:54.6121330Z * [new tag] viable/strict/1759985417 -> viable/strict/1759985417 2025-12-04T09:20:54.6121475Z * [new tag] viable/strict/1759987490 -> viable/strict/1759987490 2025-12-04T09:20:54.6121610Z * [new tag] viable/strict/1759996180 -> viable/strict/1759996180 2025-12-04T09:20:54.6121820Z * [new tag] viable/strict/1760065682 -> viable/strict/1760065682 2025-12-04T09:20:54.6121960Z * [new tag] viable/strict/1760066894 -> viable/strict/1760066894 2025-12-04T09:20:54.6122096Z * [new tag] viable/strict/1760070345 -> viable/strict/1760070345 2025-12-04T09:20:54.6122236Z * [new tag] viable/strict/1760089782 -> viable/strict/1760089782 2025-12-04T09:20:54.6122372Z * [new tag] viable/strict/1760091921 -> viable/strict/1760091921 2025-12-04T09:20:54.6122524Z * [new tag] viable/strict/1760127924 -> viable/strict/1760127924 2025-12-04T09:20:54.6122657Z * [new tag] viable/strict/1760129489 -> viable/strict/1760129489 2025-12-04T09:20:54.6122793Z * [new tag] viable/strict/1760132980 -> viable/strict/1760132980 2025-12-04T09:20:54.6122930Z * [new tag] viable/strict/1760135060 -> viable/strict/1760135060 2025-12-04T09:20:54.6123064Z * [new tag] viable/strict/1760215782 -> viable/strict/1760215782 2025-12-04T09:20:54.6123210Z * [new tag] viable/strict/1760273849 -> viable/strict/1760273849 2025-12-04T09:20:54.6123355Z * [new tag] viable/strict/1760275517 -> viable/strict/1760275517 2025-12-04T09:20:54.6123534Z * [new tag] viable/strict/1760276979 -> viable/strict/1760276979 2025-12-04T09:20:54.6123666Z * [new tag] viable/strict/1760279007 -> viable/strict/1760279007 2025-12-04T09:20:54.6123805Z * [new tag] viable/strict/1760286328 -> viable/strict/1760286328 2025-12-04T09:20:54.6123939Z * [new tag] viable/strict/1760493304 -> viable/strict/1760493304 2025-12-04T09:20:54.6124135Z * [new tag] viable/strict/1760496298 -> viable/strict/1760496298 2025-12-04T09:20:54.6129671Z * [new tag] viable/strict/1760518396 -> viable/strict/1760518396 2025-12-04T09:20:54.6130002Z * [new tag] viable/strict/1760534864 -> viable/strict/1760534864 2025-12-04T09:20:54.6130322Z * [new tag] viable/strict/1760549062 -> viable/strict/1760549062 2025-12-04T09:20:54.6130475Z * [new tag] viable/strict/1760552799 -> viable/strict/1760552799 2025-12-04T09:20:54.6130636Z * [new tag] viable/strict/1760554355 -> viable/strict/1760554355 2025-12-04T09:20:54.6130787Z * [new tag] viable/strict/1760556275 -> viable/strict/1760556275 2025-12-04T09:20:54.6131084Z * [new tag] viable/strict/1760564979 -> viable/strict/1760564979 2025-12-04T09:20:54.6131272Z * [new tag] viable/strict/1760567049 -> viable/strict/1760567049 2025-12-04T09:20:54.6131431Z * [new tag] viable/strict/1760568585 -> viable/strict/1760568585 2025-12-04T09:20:54.6131566Z * [new tag] viable/strict/1760570630 -> viable/strict/1760570630 2025-12-04T09:20:54.6131696Z * [new tag] viable/strict/1760572180 -> viable/strict/1760572180 2025-12-04T09:20:54.6131817Z * [new tag] viable/strict/1760575094 -> viable/strict/1760575094 2025-12-04T09:20:54.6131936Z * [new tag] viable/strict/1760579709 -> viable/strict/1760579709 2025-12-04T09:20:54.6132070Z * [new tag] viable/strict/1760582614 -> viable/strict/1760582614 2025-12-04T09:20:54.6132213Z * [new tag] viable/strict/1760586815 -> viable/strict/1760586815 2025-12-04T09:20:54.6132348Z * [new tag] viable/strict/1760588829 -> viable/strict/1760588829 2025-12-04T09:20:54.6138750Z * [new tag] viable/strict/1760590200 -> viable/strict/1760590200 2025-12-04T09:20:54.6139078Z * [new tag] viable/strict/1760592311 -> viable/strict/1760592311 2025-12-04T09:20:54.6139252Z * [new tag] viable/strict/1760619733 -> viable/strict/1760619733 2025-12-04T09:20:54.6139375Z * [new tag] viable/strict/1760628335 -> viable/strict/1760628335 2025-12-04T09:20:54.6139506Z * [new tag] viable/strict/1760635490 -> viable/strict/1760635490 2025-12-04T09:20:54.6139628Z * [new tag] viable/strict/1760640743 -> viable/strict/1760640743 2025-12-04T09:20:54.6139756Z * [new tag] viable/strict/1760642528 -> viable/strict/1760642528 2025-12-04T09:20:54.6141268Z * [new tag] viable/strict/1760646330 -> viable/strict/1760646330 2025-12-04T09:20:54.6141415Z * [new tag] viable/strict/1760666101 -> viable/strict/1760666101 2025-12-04T09:20:54.6141568Z * [new tag] viable/strict/1760668990 -> viable/strict/1760668990 2025-12-04T09:20:54.6142041Z * [new tag] viable/strict/1760670600 -> viable/strict/1760670600 2025-12-04T09:20:54.6142214Z * [new tag] viable/strict/1760671704 -> viable/strict/1760671704 2025-12-04T09:20:54.6142365Z * [new tag] viable/strict/1760673121 -> viable/strict/1760673121 2025-12-04T09:20:54.6142510Z * [new tag] viable/strict/1760675352 -> viable/strict/1760675352 2025-12-04T09:20:54.6142665Z * [new tag] viable/strict/1760696731 -> viable/strict/1760696731 2025-12-04T09:20:54.6143034Z * [new tag] viable/strict/1760723515 -> viable/strict/1760723515 2025-12-04T09:20:54.6145941Z * [new tag] viable/strict/1760727234 -> viable/strict/1760727234 2025-12-04T09:20:54.6146091Z * [new tag] viable/strict/1760730578 -> viable/strict/1760730578 2025-12-04T09:20:54.6146235Z * [new tag] viable/strict/1760732726 -> viable/strict/1760732726 2025-12-04T09:20:54.6146462Z * [new tag] viable/strict/1760734180 -> viable/strict/1760734180 2025-12-04T09:20:54.6146672Z * [new tag] viable/strict/1760736251 -> viable/strict/1760736251 2025-12-04T09:20:54.6146902Z * [new tag] viable/strict/1760737772 -> viable/strict/1760737772 2025-12-04T09:20:54.6147103Z * [new tag] viable/strict/1760758005 -> viable/strict/1760758005 2025-12-04T09:20:54.6147254Z * [new tag] viable/strict/1760761532 -> viable/strict/1760761532 2025-12-04T09:20:54.6150322Z * [new tag] viable/strict/1760802581 -> viable/strict/1760802581 2025-12-04T09:20:54.6150448Z * [new tag] viable/strict/1760827772 -> viable/strict/1760827772 2025-12-04T09:20:54.6150673Z * [new tag] viable/strict/1760834524 -> viable/strict/1760834524 2025-12-04T09:20:54.6150862Z * [new tag] viable/strict/1760845009 -> viable/strict/1760845009 2025-12-04T09:20:54.6151000Z * [new tag] viable/strict/1760876836 -> viable/strict/1760876836 2025-12-04T09:20:54.6151126Z * [new tag] viable/strict/1760880329 -> viable/strict/1760880329 2025-12-04T09:20:54.6151253Z * [new tag] viable/strict/1760888987 -> viable/strict/1760888987 2025-12-04T09:20:54.6151518Z * [new tag] viable/strict/1760912664 -> viable/strict/1760912664 2025-12-04T09:20:54.6151697Z * [new tag] viable/strict/1760925321 -> viable/strict/1760925321 2025-12-04T09:20:54.6155660Z * [new tag] viable/strict/1760931488 -> viable/strict/1760931488 2025-12-04T09:20:54.6155823Z * [new tag] viable/strict/1760932693 -> viable/strict/1760932693 2025-12-04T09:20:54.6155960Z * [new tag] viable/strict/1761004184 -> viable/strict/1761004184 2025-12-04T09:20:54.6156103Z * [new tag] viable/strict/1761014748 -> viable/strict/1761014748 2025-12-04T09:20:54.6156271Z * [new tag] viable/strict/1761017491 -> viable/strict/1761017491 2025-12-04T09:20:54.6156410Z * [new tag] viable/strict/1761018806 -> viable/strict/1761018806 2025-12-04T09:20:54.6156544Z * [new tag] viable/strict/1761020754 -> viable/strict/1761020754 2025-12-04T09:20:54.6156677Z * [new tag] viable/strict/1761024303 -> viable/strict/1761024303 2025-12-04T09:20:54.6156815Z * [new tag] viable/strict/1761029582 -> viable/strict/1761029582 2025-12-04T09:20:54.6156974Z * [new tag] viable/strict/1761031535 -> viable/strict/1761031535 2025-12-04T09:20:54.6157272Z * [new tag] viable/strict/1761035196 -> viable/strict/1761035196 2025-12-04T09:20:54.6157429Z * [new tag] viable/strict/1761045825 -> viable/strict/1761045825 2025-12-04T09:20:54.6158550Z * [new tag] viable/strict/1761054796 -> viable/strict/1761054796 2025-12-04T09:20:54.6159297Z * [new tag] viable/strict/1761060314 -> viable/strict/1761060314 2025-12-04T09:20:54.6159496Z * [new tag] viable/strict/1761071198 -> viable/strict/1761071198 2025-12-04T09:20:54.6160292Z * [new tag] viable/strict/1761074628 -> viable/strict/1761074628 2025-12-04T09:20:54.6160792Z * [new tag] viable/strict/1761078351 -> viable/strict/1761078351 2025-12-04T09:20:54.6161652Z * [new tag] viable/strict/1761079822 -> viable/strict/1761079822 2025-12-04T09:20:54.6162285Z * [new tag] viable/strict/1761081873 -> viable/strict/1761081873 2025-12-04T09:20:54.6162839Z * [new tag] viable/strict/1761083392 -> viable/strict/1761083392 2025-12-04T09:20:54.6166465Z * [new tag] viable/strict/1761085465 -> viable/strict/1761085465 2025-12-04T09:20:54.6167008Z * [new tag] viable/strict/1761089099 -> viable/strict/1761089099 2025-12-04T09:20:54.6167307Z * [new tag] viable/strict/1761095535 -> viable/strict/1761095535 2025-12-04T09:20:54.6167446Z * [new tag] viable/strict/1761098119 -> viable/strict/1761098119 2025-12-04T09:20:54.6167587Z * [new tag] viable/strict/1761101330 -> viable/strict/1761101330 2025-12-04T09:20:54.6167728Z * [new tag] viable/strict/1761114425 -> viable/strict/1761114425 2025-12-04T09:20:54.6167870Z * [new tag] viable/strict/1761116036 -> viable/strict/1761116036 2025-12-04T09:20:54.6168290Z * [new tag] viable/strict/1761119379 -> viable/strict/1761119379 2025-12-04T09:20:54.6172295Z * [new tag] viable/strict/1761121601 -> viable/strict/1761121601 2025-12-04T09:20:54.6172486Z * [new tag] viable/strict/1761123234 -> viable/strict/1761123234 2025-12-04T09:20:54.6172637Z * [new tag] viable/strict/1761126621 -> viable/strict/1761126621 2025-12-04T09:20:54.6172781Z * [new tag] viable/strict/1761132259 -> viable/strict/1761132259 2025-12-04T09:20:54.6172911Z * [new tag] viable/strict/1761146746 -> viable/strict/1761146746 2025-12-04T09:20:54.6173038Z * [new tag] viable/strict/1761164752 -> viable/strict/1761164752 2025-12-04T09:20:54.6173175Z * [new tag] viable/strict/1761166198 -> viable/strict/1761166198 2025-12-04T09:20:54.6173592Z * [new tag] viable/strict/1761175424 -> viable/strict/1761175424 2025-12-04T09:20:54.6174007Z * [new tag] viable/strict/1761176983 -> viable/strict/1761176983 2025-12-04T09:20:54.6174610Z * [new tag] viable/strict/1761179891 -> viable/strict/1761179891 2025-12-04T09:20:54.6178381Z * [new tag] viable/strict/1761181930 -> viable/strict/1761181930 2025-12-04T09:20:54.6178566Z * [new tag] viable/strict/1761184516 -> viable/strict/1761184516 2025-12-04T09:20:54.6178705Z * [new tag] viable/strict/1761190179 -> viable/strict/1761190179 2025-12-04T09:20:54.6179419Z * [new tag] viable/strict/1761193558 -> viable/strict/1761193558 2025-12-04T09:20:54.6179834Z * [new tag] viable/strict/1761207990 -> viable/strict/1761207990 2025-12-04T09:20:54.6179991Z * [new tag] viable/strict/1761229539 -> viable/strict/1761229539 2025-12-04T09:20:54.6180139Z * [new tag] viable/strict/1761244031 -> viable/strict/1761244031 2025-12-04T09:20:54.6180312Z * [new tag] viable/strict/1761248986 -> viable/strict/1761248986 2025-12-04T09:20:54.6180458Z * [new tag] viable/strict/1761259791 -> viable/strict/1761259791 2025-12-04T09:20:54.6180608Z * [new tag] viable/strict/1761266139 -> viable/strict/1761266139 2025-12-04T09:20:54.6183477Z * [new tag] viable/strict/1761268316 -> viable/strict/1761268316 2025-12-04T09:20:54.6183632Z * [new tag] viable/strict/1761273805 -> viable/strict/1761273805 2025-12-04T09:20:54.6183793Z * [new tag] viable/strict/1761275261 -> viable/strict/1761275261 2025-12-04T09:20:54.6184071Z * [new tag] viable/strict/1761277913 -> viable/strict/1761277913 2025-12-04T09:20:54.6184330Z * [new tag] viable/strict/1761290701 -> viable/strict/1761290701 2025-12-04T09:20:54.6184472Z * [new tag] viable/strict/1761294396 -> viable/strict/1761294396 2025-12-04T09:20:54.6184850Z * [new tag] viable/strict/1761303047 -> viable/strict/1761303047 2025-12-04T09:20:54.6186134Z * [new tag] viable/strict/1761335388 -> viable/strict/1761335388 2025-12-04T09:20:54.6186311Z * [new tag] viable/strict/1761337551 -> viable/strict/1761337551 2025-12-04T09:20:54.6187073Z * [new tag] viable/strict/1761339007 -> viable/strict/1761339007 2025-12-04T09:20:54.6187433Z * [new tag] viable/strict/1761341050 -> viable/strict/1761341050 2025-12-04T09:20:54.6191748Z * [new tag] viable/strict/1761346188 -> viable/strict/1761346188 2025-12-04T09:20:54.6192114Z * [new tag] viable/strict/1761349792 -> viable/strict/1761349792 2025-12-04T09:20:54.6192454Z * [new tag] viable/strict/1761352620 -> viable/strict/1761352620 2025-12-04T09:20:54.6192613Z * [new tag] viable/strict/1761354730 -> viable/strict/1761354730 2025-12-04T09:20:54.6192769Z * [new tag] viable/strict/1761357298 -> viable/strict/1761357298 2025-12-04T09:20:54.6192910Z * [new tag] viable/strict/1761360201 -> viable/strict/1761360201 2025-12-04T09:20:54.6193040Z * [new tag] viable/strict/1761361753 -> viable/strict/1761361753 2025-12-04T09:20:54.6193180Z * [new tag] viable/strict/1761364351 -> viable/strict/1761364351 2025-12-04T09:20:54.6193720Z * [new tag] viable/strict/1761366338 -> viable/strict/1761366338 2025-12-04T09:20:54.6194081Z * [new tag] viable/strict/1761367802 -> viable/strict/1761367802 2025-12-04T09:20:54.6194919Z * [new tag] viable/strict/1761369889 -> viable/strict/1761369889 2025-12-04T09:20:54.6195410Z * [new tag] viable/strict/1761371385 -> viable/strict/1761371385 2025-12-04T09:20:54.6196304Z * [new tag] viable/strict/1761373581 -> viable/strict/1761373581 2025-12-04T09:20:54.6196748Z * [new tag] viable/strict/1761375054 -> viable/strict/1761375054 2025-12-04T09:20:54.6197699Z * [new tag] viable/strict/1761421785 -> viable/strict/1761421785 2025-12-04T09:20:54.6198195Z * [new tag] viable/strict/1761434614 -> viable/strict/1761434614 2025-12-04T09:20:54.6199747Z * [new tag] viable/strict/1761439254 -> viable/strict/1761439254 2025-12-04T09:20:54.6200026Z * [new tag] viable/strict/1761454187 -> viable/strict/1761454187 2025-12-04T09:20:54.6201010Z * [new tag] viable/strict/1761459991 -> viable/strict/1761459991 2025-12-04T09:20:54.6204640Z * [new tag] viable/strict/1761470668 -> viable/strict/1761470668 2025-12-04T09:20:54.6209606Z * [new tag] viable/strict/1761472188 -> viable/strict/1761472188 2025-12-04T09:20:54.6214130Z * [new tag] viable/strict/1761503178 -> viable/strict/1761503178 2025-12-04T09:20:54.6216459Z * [new tag] viable/strict/1761517492 -> viable/strict/1761517492 2025-12-04T09:20:54.6216644Z * [new tag] viable/strict/1761518981 -> viable/strict/1761518981 2025-12-04T09:20:54.6216803Z * [new tag] viable/strict/1761533609 -> viable/strict/1761533609 2025-12-04T09:20:54.6216993Z * [new tag] viable/strict/1761546438 -> viable/strict/1761546438 2025-12-04T09:20:54.6217166Z * [new tag] viable/strict/1761548133 -> viable/strict/1761548133 2025-12-04T09:20:54.6217309Z * [new tag] viable/strict/1761555186 -> viable/strict/1761555186 2025-12-04T09:20:54.6217456Z * [new tag] viable/strict/1761557178 -> viable/strict/1761557178 2025-12-04T09:20:54.6217600Z * [new tag] viable/strict/1761560772 -> viable/strict/1761560772 2025-12-04T09:20:54.6217750Z * [new tag] viable/strict/1761562266 -> viable/strict/1761562266 2025-12-04T09:20:54.6218141Z * [new tag] viable/strict/1761564260 -> viable/strict/1761564260 2025-12-04T09:20:54.6218531Z * [new tag] viable/strict/1761568072 -> viable/strict/1761568072 2025-12-04T09:20:54.6218683Z * [new tag] viable/strict/1761571683 -> viable/strict/1761571683 2025-12-04T09:20:54.6218956Z * [new tag] viable/strict/1761580199 -> viable/strict/1761580199 2025-12-04T09:20:54.6219088Z * [new tag] viable/strict/1761587383 -> viable/strict/1761587383 2025-12-04T09:20:54.6219222Z * [new tag] viable/strict/1761591165 -> viable/strict/1761591165 2025-12-04T09:20:54.6219350Z * [new tag] viable/strict/1761594575 -> viable/strict/1761594575 2025-12-04T09:20:54.6219475Z * [new tag] viable/strict/1761596710 -> viable/strict/1761596710 2025-12-04T09:20:54.6219610Z * [new tag] viable/strict/1761598189 -> viable/strict/1761598189 2025-12-04T09:20:54.6219748Z * [new tag] viable/strict/1761600254 -> viable/strict/1761600254 2025-12-04T09:20:54.6219883Z * [new tag] viable/strict/1761603879 -> viable/strict/1761603879 2025-12-04T09:20:54.6220006Z * [new tag] viable/strict/1761605429 -> viable/strict/1761605429 2025-12-04T09:20:54.6220133Z * [new tag] viable/strict/1761607468 -> viable/strict/1761607468 2025-12-04T09:20:54.6220266Z * [new tag] viable/strict/1761608983 -> viable/strict/1761608983 2025-12-04T09:20:54.6220388Z * [new tag] viable/strict/1761611846 -> viable/strict/1761611846 2025-12-04T09:20:54.6220518Z * [new tag] viable/strict/1761613922 -> viable/strict/1761613922 2025-12-04T09:20:54.6221186Z * [new tag] viable/strict/1761616504 -> viable/strict/1761616504 2025-12-04T09:20:54.6221782Z * [new tag] viable/strict/1761619599 -> viable/strict/1761619599 2025-12-04T09:20:54.6221971Z * [new tag] viable/strict/1761686693 -> viable/strict/1761686693 2025-12-04T09:20:54.6222107Z * [new tag] viable/strict/1761688179 -> viable/strict/1761688179 2025-12-04T09:20:54.6222238Z * [new tag] viable/strict/1761691973 -> viable/strict/1761691973 2025-12-04T09:20:54.6222383Z * [new tag] viable/strict/1761693884 -> viable/strict/1761693884 2025-12-04T09:20:54.6222515Z * [new tag] viable/strict/1761695389 -> viable/strict/1761695389 2025-12-04T09:20:54.6225072Z * [new tag] viable/strict/1761698408 -> viable/strict/1761698408 2025-12-04T09:20:54.6225285Z * [new tag] viable/strict/1761702931 -> viable/strict/1761702931 2025-12-04T09:20:54.6225458Z * [new tag] viable/strict/1761706307 -> viable/strict/1761706307 2025-12-04T09:20:54.6225624Z * [new tag] viable/strict/1761709065 -> viable/strict/1761709065 2025-12-04T09:20:54.6225820Z * [new tag] viable/strict/1761710285 -> viable/strict/1761710285 2025-12-04T09:20:54.6229381Z * [new tag] viable/strict/1761711983 -> viable/strict/1761711983 2025-12-04T09:20:54.6229892Z * [new tag] viable/strict/1761713514 -> viable/strict/1761713514 2025-12-04T09:20:54.6230047Z * [new tag] viable/strict/1761715523 -> viable/strict/1761715523 2025-12-04T09:20:54.6230331Z * [new tag] viable/strict/1761727973 -> viable/strict/1761727973 2025-12-04T09:20:54.6230476Z * [new tag] viable/strict/1761751558 -> viable/strict/1761751558 2025-12-04T09:20:54.6230607Z * [new tag] viable/strict/1761755187 -> viable/strict/1761755187 2025-12-04T09:20:54.6236775Z * [new tag] viable/strict/1761756826 -> viable/strict/1761756826 2025-12-04T09:20:54.6239735Z * [new tag] viable/strict/1761769551 -> viable/strict/1761769551 2025-12-04T09:20:54.6240135Z * [new tag] viable/strict/1761771032 -> viable/strict/1761771032 2025-12-04T09:20:54.6240285Z * [new tag] viable/strict/1761773101 -> viable/strict/1761773101 2025-12-04T09:20:54.6240417Z * [new tag] viable/strict/1761781792 -> viable/strict/1761781792 2025-12-04T09:20:54.6240609Z * [new tag] viable/strict/1761784788 -> viable/strict/1761784788 2025-12-04T09:20:54.6240752Z * [new tag] viable/strict/1761786740 -> viable/strict/1761786740 2025-12-04T09:20:54.6240883Z * [new tag] viable/strict/1761789332 -> viable/strict/1761789332 2025-12-04T09:20:54.6241022Z * [new tag] viable/strict/1761792569 -> viable/strict/1761792569 2025-12-04T09:20:54.6241149Z * [new tag] viable/strict/1761795289 -> viable/strict/1761795289 2025-12-04T09:20:54.6241277Z * [new tag] viable/strict/1761798345 -> viable/strict/1761798345 2025-12-04T09:20:54.6241421Z * [new tag] viable/strict/1761799827 -> viable/strict/1761799827 2025-12-04T09:20:54.6241551Z * [new tag] viable/strict/1761805604 -> viable/strict/1761805604 2025-12-04T09:20:54.6242028Z * [new tag] viable/strict/1761807202 -> viable/strict/1761807202 2025-12-04T09:20:54.6242190Z * [new tag] viable/strict/1761809094 -> viable/strict/1761809094 2025-12-04T09:20:54.6242319Z * [new tag] viable/strict/1761810576 -> viable/strict/1761810576 2025-12-04T09:20:54.6242470Z * [new tag] viable/strict/1761812771 -> viable/strict/1761812771 2025-12-04T09:20:54.6242619Z * [new tag] viable/strict/1761814363 -> viable/strict/1761814363 2025-12-04T09:20:54.6247686Z * [new tag] viable/strict/1761857410 -> viable/strict/1761857410 2025-12-04T09:20:54.6248138Z * [new tag] viable/strict/1761860985 -> viable/strict/1761860985 2025-12-04T09:20:54.6248321Z * [new tag] viable/strict/1761863094 -> viable/strict/1761863094 2025-12-04T09:20:54.6248473Z * [new tag] viable/strict/1761864590 -> viable/strict/1761864590 2025-12-04T09:20:54.6248621Z * [new tag] viable/strict/1761866675 -> viable/strict/1761866675 2025-12-04T09:20:54.6248786Z * [new tag] viable/strict/1761868178 -> viable/strict/1761868178 2025-12-04T09:20:54.6248923Z * [new tag] viable/strict/1761871111 -> viable/strict/1761871111 2025-12-04T09:20:54.6249065Z * [new tag] viable/strict/1761873126 -> viable/strict/1761873126 2025-12-04T09:20:54.6249202Z * [new tag] viable/strict/1761875714 -> viable/strict/1761875714 2025-12-04T09:20:54.6252895Z * [new tag] viable/strict/1761878924 -> viable/strict/1761878924 2025-12-04T09:20:54.6253158Z * [new tag] viable/strict/1761881727 -> viable/strict/1761881727 2025-12-04T09:20:54.6258520Z * [new tag] viable/strict/1761882959 -> viable/strict/1761882959 2025-12-04T09:20:54.6265160Z * [new tag] viable/strict/1761886268 -> viable/strict/1761886268 2025-12-04T09:20:54.6265487Z * [new tag] viable/strict/1761893641 -> viable/strict/1761893641 2025-12-04T09:20:54.6265657Z * [new tag] viable/strict/1761931517 -> viable/strict/1761931517 2025-12-04T09:20:54.6265879Z * [new tag] viable/strict/1761933080 -> viable/strict/1761933080 2025-12-04T09:20:54.6266024Z * [new tag] viable/strict/1761935217 -> viable/strict/1761935217 2025-12-04T09:20:54.6266163Z * [new tag] viable/strict/1761938533 -> viable/strict/1761938533 2025-12-04T09:20:54.6266313Z * [new tag] viable/strict/1761940184 -> viable/strict/1761940184 2025-12-04T09:20:54.6266457Z * [new tag] viable/strict/1761942338 -> viable/strict/1761942338 2025-12-04T09:20:54.6267203Z * [new tag] viable/strict/1761946100 -> viable/strict/1761946100 2025-12-04T09:20:54.6267392Z * [new tag] viable/strict/1761947374 -> viable/strict/1761947374 2025-12-04T09:20:54.6267530Z * [new tag] viable/strict/1761950978 -> viable/strict/1761950978 2025-12-04T09:20:54.6267828Z * [new tag] viable/strict/1761957727 -> viable/strict/1761957727 2025-12-04T09:20:54.6267964Z * [new tag] viable/strict/1761959532 -> viable/strict/1761959532 2025-12-04T09:20:54.6268099Z * [new tag] viable/strict/1761965366 -> viable/strict/1761965366 2025-12-04T09:20:54.6268240Z * [new tag] viable/strict/1761968066 -> viable/strict/1761968066 2025-12-04T09:20:54.6268369Z * [new tag] viable/strict/1761969322 -> viable/strict/1761969322 2025-12-04T09:20:54.6268522Z * [new tag] viable/strict/1761974723 -> viable/strict/1761974723 2025-12-04T09:20:54.6268659Z * [new tag] viable/strict/1761981837 -> viable/strict/1761981837 2025-12-04T09:20:54.6268786Z * [new tag] viable/strict/1761985546 -> viable/strict/1761985546 2025-12-04T09:20:54.6268922Z * [new tag] viable/strict/1761987030 -> viable/strict/1761987030 2025-12-04T09:20:54.6269054Z * [new tag] viable/strict/1762003554 -> viable/strict/1762003554 2025-12-04T09:20:54.6269180Z * [new tag] viable/strict/1762021560 -> viable/strict/1762021560 2025-12-04T09:20:54.6269317Z * [new tag] viable/strict/1762032190 -> viable/strict/1762032190 2025-12-04T09:20:54.6269444Z * [new tag] viable/strict/1762040981 -> viable/strict/1762040981 2025-12-04T09:20:54.6269584Z * [new tag] viable/strict/1762048525 -> viable/strict/1762048525 2025-12-04T09:20:54.6269706Z * [new tag] viable/strict/1762104223 -> viable/strict/1762104223 2025-12-04T09:20:54.6269834Z * [new tag] viable/strict/1762105778 -> viable/strict/1762105778 2025-12-04T09:20:54.6269965Z * [new tag] viable/strict/1762115109 -> viable/strict/1762115109 2025-12-04T09:20:54.6270088Z * [new tag] viable/strict/1762125840 -> viable/strict/1762125840 2025-12-04T09:20:54.6270221Z * [new tag] viable/strict/1762127377 -> viable/strict/1762127377 2025-12-04T09:20:54.6270347Z * [new tag] viable/strict/1762134925 -> viable/strict/1762134925 2025-12-04T09:20:54.6270472Z * [new tag] viable/strict/1762138338 -> viable/strict/1762138338 2025-12-04T09:20:54.6270798Z * [new tag] viable/strict/1762148993 -> viable/strict/1762148993 2025-12-04T09:20:54.6271135Z * [new tag] viable/strict/1762152871 -> viable/strict/1762152871 2025-12-04T09:20:54.6271290Z * [new tag] viable/strict/1762156183 -> viable/strict/1762156183 2025-12-04T09:20:54.6271424Z * [new tag] viable/strict/1762163457 -> viable/strict/1762163457 2025-12-04T09:20:54.6271712Z * [new tag] viable/strict/1762165569 -> viable/strict/1762165569 2025-12-04T09:20:54.6273190Z * [new tag] viable/strict/1762169035 -> viable/strict/1762169035 2025-12-04T09:20:54.6273519Z * [new tag] viable/strict/1762174936 -> viable/strict/1762174936 2025-12-04T09:20:54.6278359Z * [new tag] viable/strict/1762194412 -> viable/strict/1762194412 2025-12-04T09:20:54.6278639Z * [new tag] viable/strict/1762195876 -> viable/strict/1762195876 2025-12-04T09:20:54.6278912Z * [new tag] viable/strict/1762197788 -> viable/strict/1762197788 2025-12-04T09:20:54.6279048Z * [new tag] viable/strict/1762199389 -> viable/strict/1762199389 2025-12-04T09:20:54.6279183Z * [new tag] viable/strict/1762206585 -> viable/strict/1762206585 2025-12-04T09:20:54.6279622Z * [new tag] viable/strict/1762210184 -> viable/strict/1762210184 2025-12-04T09:20:54.6279832Z * [new tag] viable/strict/1762218736 -> viable/strict/1762218736 2025-12-04T09:20:54.6279985Z * [new tag] viable/strict/1762224529 -> viable/strict/1762224529 2025-12-04T09:20:54.6280177Z * [new tag] viable/strict/1762227253 -> viable/strict/1762227253 2025-12-04T09:20:54.6280315Z * [new tag] viable/strict/1762228515 -> viable/strict/1762228515 2025-12-04T09:20:54.6280458Z * [new tag] viable/strict/1762230349 -> viable/strict/1762230349 2025-12-04T09:20:54.6280598Z * [new tag] viable/strict/1762231859 -> viable/strict/1762231859 2025-12-04T09:20:54.6280744Z * [new tag] viable/strict/1762233925 -> viable/strict/1762233925 2025-12-04T09:20:54.6280892Z * [new tag] viable/strict/1762237630 -> viable/strict/1762237630 2025-12-04T09:20:54.6281027Z * [new tag] viable/strict/1762253522 -> viable/strict/1762253522 2025-12-04T09:20:54.6281535Z * [new tag] viable/strict/1762278588 -> viable/strict/1762278588 2025-12-04T09:20:54.6282324Z * [new tag] viable/strict/1762284203 -> viable/strict/1762284203 2025-12-04T09:20:54.6283037Z * [new tag] viable/strict/1762289446 -> viable/strict/1762289446 2025-12-04T09:20:54.6286501Z * [new tag] viable/strict/1762291515 -> viable/strict/1762291515 2025-12-04T09:20:54.6286818Z * [new tag] viable/strict/1762295100 -> viable/strict/1762295100 2025-12-04T09:20:54.6286980Z * [new tag] viable/strict/1762296590 -> viable/strict/1762296590 2025-12-04T09:20:54.6287107Z * [new tag] viable/strict/1762300179 -> viable/strict/1762300179 2025-12-04T09:20:54.6287242Z * [new tag] viable/strict/1762303207 -> viable/strict/1762303207 2025-12-04T09:20:54.6287530Z * [new tag] viable/strict/1762386584 -> viable/strict/1762386584 2025-12-04T09:20:54.6287677Z * [new tag] viable/strict/1762391537 -> viable/strict/1762391537 2025-12-04T09:20:54.6287911Z * [new tag] viable/strict/1762394119 -> viable/strict/1762394119 2025-12-04T09:20:54.6292138Z * [new tag] viable/strict/1762397437 -> viable/strict/1762397437 2025-12-04T09:20:54.6292429Z * [new tag] viable/strict/1762400256 -> viable/strict/1762400256 2025-12-04T09:20:54.6298998Z * [new tag] viable/strict/1762401469 -> viable/strict/1762401469 2025-12-04T09:20:54.6304540Z * [new tag] viable/strict/1762408195 -> viable/strict/1762408195 2025-12-04T09:20:54.6309149Z * [new tag] viable/strict/1762410411 -> viable/strict/1762410411 2025-12-04T09:20:54.6314119Z * [new tag] viable/strict/1762417613 -> viable/strict/1762417613 2025-12-04T09:20:54.6317826Z * [new tag] viable/strict/1762419198 -> viable/strict/1762419198 2025-12-04T09:20:54.6320114Z * [new tag] viable/strict/1762422656 -> viable/strict/1762422656 2025-12-04T09:20:54.6320413Z * [new tag] viable/strict/1762424746 -> viable/strict/1762424746 2025-12-04T09:20:54.6320682Z * [new tag] viable/strict/1762446386 -> viable/strict/1762446386 2025-12-04T09:20:54.6320877Z * [new tag] viable/strict/1762449912 -> viable/strict/1762449912 2025-12-04T09:20:54.6321049Z * [new tag] viable/strict/1762457031 -> viable/strict/1762457031 2025-12-04T09:20:54.6321187Z * [new tag] viable/strict/1762462441 -> viable/strict/1762462441 2025-12-04T09:20:54.6321336Z * [new tag] viable/strict/1762467909 -> viable/strict/1762467909 2025-12-04T09:20:54.6321479Z * [new tag] viable/strict/1762471493 -> viable/strict/1762471493 2025-12-04T09:20:54.6322066Z * [new tag] viable/strict/1762475990 -> viable/strict/1762475990 2025-12-04T09:20:54.6322602Z * [new tag] viable/strict/1762477933 -> viable/strict/1762477933 2025-12-04T09:20:54.6322983Z * [new tag] viable/strict/1762491053 -> viable/strict/1762491053 2025-12-04T09:20:54.6323286Z * [new tag] viable/strict/1762493118 -> viable/strict/1762493118 2025-12-04T09:20:54.6323426Z * [new tag] viable/strict/1762498442 -> viable/strict/1762498442 2025-12-04T09:20:54.6323596Z * [new tag] viable/strict/1762501778 -> viable/strict/1762501778 2025-12-04T09:20:54.6323883Z * [new tag] viable/strict/1762504001 -> viable/strict/1762504001 2025-12-04T09:20:54.6324039Z * [new tag] viable/strict/1762505583 -> viable/strict/1762505583 2025-12-04T09:20:54.6324283Z * [new tag] viable/strict/1762507523 -> viable/strict/1762507523 2025-12-04T09:20:54.6324430Z * [new tag] viable/strict/1762511140 -> viable/strict/1762511140 2025-12-04T09:20:54.6324642Z * [new tag] viable/strict/1762512632 -> viable/strict/1762512632 2025-12-04T09:20:54.6324826Z * [new tag] viable/strict/1762520467 -> viable/strict/1762520467 2025-12-04T09:20:54.6324972Z * [new tag] viable/strict/1762522016 -> viable/strict/1762522016 2025-12-04T09:20:54.6325107Z * [new tag] viable/strict/1762530591 -> viable/strict/1762530591 2025-12-04T09:20:54.6325267Z * [new tag] viable/strict/1762543405 -> viable/strict/1762543405 2025-12-04T09:20:54.6325538Z * [new tag] viable/strict/1762544998 -> viable/strict/1762544998 2025-12-04T09:20:54.6325750Z * [new tag] viable/strict/1762552182 -> viable/strict/1762552182 2025-12-04T09:20:54.6325968Z * [new tag] viable/strict/1762554297 -> viable/strict/1762554297 2025-12-04T09:20:54.6326113Z * [new tag] viable/strict/1762559381 -> viable/strict/1762559381 2025-12-04T09:20:54.6326279Z * [new tag] viable/strict/1762562222 -> viable/strict/1762562222 2025-12-04T09:20:54.6326495Z * [new tag] viable/strict/1762564319 -> viable/strict/1762564319 2025-12-04T09:20:54.6327073Z * [new tag] viable/strict/1762566904 -> viable/strict/1762566904 2025-12-04T09:20:54.6327367Z * [new tag] viable/strict/1762569781 -> viable/strict/1762569781 2025-12-04T09:20:54.6327590Z * [new tag] viable/strict/1762575940 -> viable/strict/1762575940 2025-12-04T09:20:54.6327750Z * [new tag] viable/strict/1762580974 -> viable/strict/1762580974 2025-12-04T09:20:54.6327889Z * [new tag] viable/strict/1762583185 -> viable/strict/1762583185 2025-12-04T09:20:54.6328094Z * [new tag] viable/strict/1762586647 -> viable/strict/1762586647 2025-12-04T09:20:54.6333712Z * [new tag] viable/strict/1762588183 -> viable/strict/1762588183 2025-12-04T09:20:54.6336031Z * [new tag] viable/strict/1762593886 -> viable/strict/1762593886 2025-12-04T09:20:54.6343278Z * [new tag] viable/strict/1762650743 -> viable/strict/1762650743 2025-12-04T09:20:54.6343608Z * [new tag] viable/strict/1762653328 -> viable/strict/1762653328 2025-12-04T09:20:54.6343771Z * [new tag] viable/strict/1762659342 -> viable/strict/1762659342 2025-12-04T09:20:54.6343916Z * [new tag] viable/strict/1762662360 -> viable/strict/1762662360 2025-12-04T09:20:54.6344091Z * [new tag] viable/strict/1762667377 -> viable/strict/1762667377 2025-12-04T09:20:54.6344232Z * [new tag] viable/strict/1762671090 -> viable/strict/1762671090 2025-12-04T09:20:54.6344380Z * [new tag] viable/strict/1762680284 -> viable/strict/1762680284 2025-12-04T09:20:54.6344738Z * [new tag] viable/strict/1762683900 -> viable/strict/1762683900 2025-12-04T09:20:54.6345327Z * [new tag] viable/strict/1762705541 -> viable/strict/1762705541 2025-12-04T09:20:54.6345484Z * [new tag] viable/strict/1762709004 -> viable/strict/1762709004 2025-12-04T09:20:54.6345854Z * [new tag] viable/strict/1762746004 -> viable/strict/1762746004 2025-12-04T09:20:54.6346016Z * [new tag] viable/strict/1762748799 -> viable/strict/1762748799 2025-12-04T09:20:54.6346165Z * [new tag] viable/strict/1762759504 -> viable/strict/1762759504 2025-12-04T09:20:54.6346314Z * [new tag] viable/strict/1762760973 -> viable/strict/1762760973 2025-12-04T09:20:54.6346457Z * [new tag] viable/strict/1762775374 -> viable/strict/1762775374 2025-12-04T09:20:54.6346596Z * [new tag] viable/strict/1762777661 -> viable/strict/1762777661 2025-12-04T09:20:54.6346741Z * [new tag] viable/strict/1762779774 -> viable/strict/1762779774 2025-12-04T09:20:54.6346881Z * [new tag] viable/strict/1762781259 -> viable/strict/1762781259 2025-12-04T09:20:54.6347030Z * [new tag] viable/strict/1762793628 -> viable/strict/1762793628 2025-12-04T09:20:54.6347175Z * [new tag] viable/strict/1762800711 -> viable/strict/1762800711 2025-12-04T09:20:54.6347303Z * [new tag] viable/strict/1762809894 -> viable/strict/1762809894 2025-12-04T09:20:54.6347452Z * [new tag] viable/strict/1762811384 -> viable/strict/1762811384 2025-12-04T09:20:54.6347592Z * [new tag] viable/strict/1762813841 -> viable/strict/1762813841 2025-12-04T09:20:54.6347726Z * [new tag] viable/strict/1762815047 -> viable/strict/1762815047 2025-12-04T09:20:54.6347874Z * [new tag] viable/strict/1762817094 -> viable/strict/1762817094 2025-12-04T09:20:54.6348002Z * [new tag] viable/strict/1762818582 -> viable/strict/1762818582 2025-12-04T09:20:54.6348145Z * [new tag] viable/strict/1762821623 -> viable/strict/1762821623 2025-12-04T09:20:54.6348283Z * [new tag] viable/strict/1762823531 -> viable/strict/1762823531 2025-12-04T09:20:54.6348422Z * [new tag] viable/strict/1762849583 -> viable/strict/1762849583 2025-12-04T09:20:54.6348552Z * [new tag] viable/strict/1762851200 -> viable/strict/1762851200 2025-12-04T09:20:54.6348690Z * [new tag] viable/strict/1762854603 -> viable/strict/1762854603 2025-12-04T09:20:54.6348834Z * [new tag] viable/strict/1762858276 -> viable/strict/1762858276 2025-12-04T09:20:54.6348972Z * [new tag] viable/strict/1762860891 -> viable/strict/1762860891 2025-12-04T09:20:54.6349111Z * [new tag] viable/strict/1762866174 -> viable/strict/1762866174 2025-12-04T09:20:54.6349245Z * [new tag] viable/strict/1762867653 -> viable/strict/1762867653 2025-12-04T09:20:54.6349405Z * [new tag] viable/strict/1762872669 -> viable/strict/1762872669 2025-12-04T09:20:54.6349540Z * [new tag] viable/strict/1762878380 -> viable/strict/1762878380 2025-12-04T09:20:54.6349667Z * [new tag] viable/strict/1762889003 -> viable/strict/1762889003 2025-12-04T09:20:54.6349785Z * [new tag] viable/strict/1762890589 -> viable/strict/1762890589 2025-12-04T09:20:54.6349910Z * [new tag] viable/strict/1762892743 -> viable/strict/1762892743 2025-12-04T09:20:54.6350028Z * [new tag] viable/strict/1762894271 -> viable/strict/1762894271 2025-12-04T09:20:54.6350145Z * [new tag] viable/strict/1762896287 -> viable/strict/1762896287 2025-12-04T09:20:54.6350269Z * [new tag] viable/strict/1762915871 -> viable/strict/1762915871 2025-12-04T09:20:54.6350435Z * [new tag] viable/strict/1762918569 -> viable/strict/1762918569 2025-12-04T09:20:54.6350568Z * [new tag] viable/strict/1762919776 -> viable/strict/1762919776 2025-12-04T09:20:54.6350688Z * [new tag] viable/strict/1762923072 -> viable/strict/1762923072 2025-12-04T09:20:54.6350859Z * [new tag] viable/strict/1762928826 -> viable/strict/1762928826 2025-12-04T09:20:54.6350992Z * [new tag] viable/strict/1762930451 -> viable/strict/1762930451 2025-12-04T09:20:54.6351130Z * [new tag] viable/strict/1762933780 -> viable/strict/1762933780 2025-12-04T09:20:54.6351274Z * [new tag] viable/strict/1762937638 -> viable/strict/1762937638 2025-12-04T09:20:54.6354739Z * [new tag] viable/strict/1762939545 -> viable/strict/1762939545 2025-12-04T09:20:54.6354890Z * [new tag] viable/strict/1762962692 -> viable/strict/1762962692 2025-12-04T09:20:54.6355071Z * [new tag] viable/strict/1762979143 -> viable/strict/1762979143 2025-12-04T09:20:54.6355223Z * [new tag] viable/strict/1762984188 -> viable/strict/1762984188 2025-12-04T09:20:54.6355382Z * [new tag] viable/strict/1762986306 -> viable/strict/1762986306 2025-12-04T09:20:54.6355526Z * [new tag] viable/strict/1762989903 -> viable/strict/1762989903 2025-12-04T09:20:54.6355667Z * [new tag] viable/strict/1762991377 -> viable/strict/1762991377 2025-12-04T09:20:54.6355805Z * [new tag] viable/strict/1762998921 -> viable/strict/1762998921 2025-12-04T09:20:54.6355942Z * [new tag] viable/strict/1763002287 -> viable/strict/1763002287 2025-12-04T09:20:54.6356556Z * [new tag] viable/strict/1763016840 -> viable/strict/1763016840 2025-12-04T09:20:54.6357005Z * [new tag] viable/strict/1763020180 -> viable/strict/1763020180 2025-12-04T09:20:54.6357191Z * [new tag] viable/strict/1763027421 -> viable/strict/1763027421 2025-12-04T09:20:54.6357326Z * [new tag] viable/strict/1763031120 -> viable/strict/1763031120 2025-12-04T09:20:54.6357468Z * [new tag] viable/strict/1763036861 -> viable/strict/1763036861 2025-12-04T09:20:54.6357620Z * [new tag] viable/strict/1763038993 -> viable/strict/1763038993 2025-12-04T09:20:54.6357753Z * [new tag] viable/strict/1763054703 -> viable/strict/1763054703 2025-12-04T09:20:54.6357890Z * [new tag] viable/strict/1763067061 -> viable/strict/1763067061 2025-12-04T09:20:54.6358059Z * [new tag] viable/strict/1763070847 -> viable/strict/1763070847 2025-12-04T09:20:54.6359361Z * [new tag] viable/strict/1763072706 -> viable/strict/1763072706 2025-12-04T09:20:54.6359884Z * [new tag] viable/strict/1763076302 -> viable/strict/1763076302 2025-12-04T09:20:54.6360580Z * [new tag] viable/strict/1763080816 -> viable/strict/1763080816 2025-12-04T09:20:54.6361051Z * [new tag] viable/strict/1763082732 -> viable/strict/1763082732 2025-12-04T09:20:54.6362165Z * [new tag] viable/strict/1763085329 -> viable/strict/1763085329 2025-12-04T09:20:54.6362434Z * [new tag] viable/strict/1763088623 -> viable/strict/1763088623 2025-12-04T09:20:54.6365775Z * [new tag] viable/strict/1763091402 -> viable/strict/1763091402 2025-12-04T09:20:54.6366075Z * [new tag] viable/strict/1763092602 -> viable/strict/1763092602 2025-12-04T09:20:54.6366410Z * [new tag] viable/strict/1763094355 -> viable/strict/1763094355 2025-12-04T09:20:54.6366558Z * [new tag] viable/strict/1763099390 -> viable/strict/1763099390 2025-12-04T09:20:54.6366691Z * [new tag] viable/strict/1763101608 -> viable/strict/1763101608 2025-12-04T09:20:54.6366976Z * [new tag] viable/strict/1763105102 -> viable/strict/1763105102 2025-12-04T09:20:54.6367437Z * [new tag] viable/strict/1763112347 -> viable/strict/1763112347 2025-12-04T09:20:54.6371954Z * [new tag] viable/strict/1763119471 -> viable/strict/1763119471 2025-12-04T09:20:54.6372425Z * [new tag] viable/strict/1763126835 -> viable/strict/1763126835 2025-12-04T09:20:54.6372689Z * [new tag] viable/strict/1763149779 -> viable/strict/1763149779 2025-12-04T09:20:54.6373002Z * [new tag] viable/strict/1763164178 -> viable/strict/1763164178 2025-12-04T09:20:54.6373262Z * [new tag] viable/strict/1763167104 -> viable/strict/1763167104 2025-12-04T09:20:54.6373405Z * [new tag] viable/strict/1763169132 -> viable/strict/1763169132 2025-12-04T09:20:54.6373546Z * [new tag] viable/strict/1763171708 -> viable/strict/1763171708 2025-12-04T09:20:54.6373683Z * [new tag] viable/strict/1763174759 -> viable/strict/1763174759 2025-12-04T09:20:54.6373822Z * [new tag] viable/strict/1763180744 -> viable/strict/1763180744 2025-12-04T09:20:54.6374192Z * [new tag] viable/strict/1763182227 -> viable/strict/1763182227 2025-12-04T09:20:54.6377502Z * [new tag] viable/strict/1763184309 -> viable/strict/1763184309 2025-12-04T09:20:54.6378050Z * [new tag] viable/strict/1763187991 -> viable/strict/1763187991 2025-12-04T09:20:54.6378203Z * [new tag] viable/strict/1763191445 -> viable/strict/1763191445 2025-12-04T09:20:54.6378342Z * [new tag] viable/strict/1763195152 -> viable/strict/1763195152 2025-12-04T09:20:54.6378467Z * [new tag] viable/strict/1763205769 -> viable/strict/1763205769 2025-12-04T09:20:54.6383644Z * [new tag] viable/strict/1763246990 -> viable/strict/1763246990 2025-12-04T09:20:54.6389231Z * [new tag] viable/strict/1763261578 -> viable/strict/1763261578 2025-12-04T09:20:54.6389615Z * [new tag] viable/strict/1763286573 -> viable/strict/1763286573 2025-12-04T09:20:54.6389764Z * [new tag] viable/strict/1763292167 -> viable/strict/1763292167 2025-12-04T09:20:54.6389983Z * [new tag] viable/strict/1763333386 -> viable/strict/1763333386 2025-12-04T09:20:54.6390140Z * [new tag] viable/strict/1763340082 -> viable/strict/1763340082 2025-12-04T09:20:54.6390286Z * [new tag] viable/strict/1763364324 -> viable/strict/1763364324 2025-12-04T09:20:54.6390432Z * [new tag] viable/strict/1763371569 -> viable/strict/1763371569 2025-12-04T09:20:54.6390588Z * [new tag] viable/strict/1763373067 -> viable/strict/1763373067 2025-12-04T09:20:54.6390729Z * [new tag] viable/strict/1763375157 -> viable/strict/1763375157 2025-12-04T09:20:54.6390886Z * [new tag] viable/strict/1763382462 -> viable/strict/1763382462 2025-12-04T09:20:54.6391032Z * [new tag] viable/strict/1763394661 -> viable/strict/1763394661 2025-12-04T09:20:54.6391737Z * [new tag] viable/strict/1763396797 -> viable/strict/1763396797 2025-12-04T09:20:54.6391964Z * [new tag] viable/strict/1763398542 -> viable/strict/1763398542 2025-12-04T09:20:54.6392096Z * [new tag] viable/strict/1763401807 -> viable/strict/1763401807 2025-12-04T09:20:54.6392233Z * [new tag] viable/strict/1763414698 -> viable/strict/1763414698 2025-12-04T09:20:54.6392369Z * [new tag] viable/strict/1763419807 -> viable/strict/1763419807 2025-12-04T09:20:54.6392499Z * [new tag] viable/strict/1763426369 -> viable/strict/1763426369 2025-12-04T09:20:54.6392637Z * [new tag] viable/strict/1763428331 -> viable/strict/1763428331 2025-12-04T09:20:54.6398106Z * [new tag] viable/strict/1763430922 -> viable/strict/1763430922 2025-12-04T09:20:54.6398672Z * [new tag] viable/strict/1763434184 -> viable/strict/1763434184 2025-12-04T09:20:54.6398830Z * [new tag] viable/strict/1763439973 -> viable/strict/1763439973 2025-12-04T09:20:54.6399161Z * [new tag] viable/strict/1763444995 -> viable/strict/1763444995 2025-12-04T09:20:54.6399316Z * [new tag] viable/strict/1763447206 -> viable/strict/1763447206 2025-12-04T09:20:54.6399448Z * [new tag] viable/strict/1763448826 -> viable/strict/1763448826 2025-12-04T09:20:54.6399585Z * [new tag] viable/strict/1763450717 -> viable/strict/1763450717 2025-12-04T09:20:54.6399712Z * [new tag] viable/strict/1763452183 -> viable/strict/1763452183 2025-12-04T09:20:54.6399842Z * [new tag] viable/strict/1763457945 -> viable/strict/1763457945 2025-12-04T09:20:54.6399990Z * [new tag] viable/strict/1763459439 -> viable/strict/1763459439 2025-12-04T09:20:54.6400117Z * [new tag] viable/strict/1763461556 -> viable/strict/1763461556 2025-12-04T09:20:54.6400260Z * [new tag] viable/strict/1763463103 -> viable/strict/1763463103 2025-12-04T09:20:54.6400399Z * [new tag] viable/strict/1763465100 -> viable/strict/1763465100 2025-12-04T09:20:54.6400566Z * [new tag] viable/strict/1763468866 -> viable/strict/1763468866 2025-12-04T09:20:54.6401450Z * [new tag] viable/strict/1763493823 -> viable/strict/1763493823 2025-12-04T09:20:54.6401640Z * [new tag] viable/strict/1763496249 -> viable/strict/1763496249 2025-12-04T09:20:54.6402908Z * [new tag] viable/strict/1763502620 -> viable/strict/1763502620 2025-12-04T09:20:54.6407513Z * [new tag] viable/strict/1763504715 -> viable/strict/1763504715 2025-12-04T09:20:54.6407713Z * [new tag] viable/strict/1763506208 -> viable/strict/1763506208 2025-12-04T09:20:54.6407851Z * [new tag] viable/strict/1763520590 -> viable/strict/1763520590 2025-12-04T09:20:54.6407989Z * [new tag] viable/strict/1763523357 -> viable/strict/1763523357 2025-12-04T09:20:54.6408157Z * [new tag] viable/strict/1763529922 -> viable/strict/1763529922 2025-12-04T09:20:54.6408286Z * [new tag] viable/strict/1763531408 -> viable/strict/1763531408 2025-12-04T09:20:54.6408424Z * [new tag] viable/strict/1763533622 -> viable/strict/1763533622 2025-12-04T09:20:54.6408730Z * [new tag] viable/strict/1763538576 -> viable/strict/1763538576 2025-12-04T09:20:54.6408985Z * [new tag] viable/strict/1763545823 -> viable/strict/1763545823 2025-12-04T09:20:54.6409793Z * [new tag] viable/strict/1763547951 -> viable/strict/1763547951 2025-12-04T09:20:54.6410277Z * [new tag] viable/strict/1763551477 -> viable/strict/1763551477 2025-12-04T09:20:54.6413379Z * [new tag] viable/strict/1763552982 -> viable/strict/1763552982 2025-12-04T09:20:54.6413688Z * [new tag] viable/strict/1763594698 -> viable/strict/1763594698 2025-12-04T09:20:54.6413867Z * [new tag] viable/strict/1763596178 -> viable/strict/1763596178 2025-12-04T09:20:54.6414001Z * [new tag] viable/strict/1763599155 -> viable/strict/1763599155 2025-12-04T09:20:54.6418926Z * [new tag] viable/strict/1763603717 -> viable/strict/1763603717 2025-12-04T09:20:54.6419239Z * [new tag] viable/strict/1763606923 -> viable/strict/1763606923 2025-12-04T09:20:54.6419398Z * [new tag] viable/strict/1763609715 -> viable/strict/1763609715 2025-12-04T09:20:54.6419544Z * [new tag] viable/strict/1763612757 -> viable/strict/1763612757 2025-12-04T09:20:54.6419970Z * [new tag] viable/strict/1763616325 -> viable/strict/1763616325 2025-12-04T09:20:54.6420182Z * [new tag] viable/strict/1763623509 -> viable/strict/1763623509 2025-12-04T09:20:54.6420834Z * [new tag] viable/strict/1763624984 -> viable/strict/1763624984 2025-12-04T09:20:54.6421304Z * [new tag] viable/strict/1763628796 -> viable/strict/1763628796 2025-12-04T09:20:54.6421724Z * [new tag] viable/strict/1763634343 -> viable/strict/1763634343 2025-12-04T09:20:54.6421902Z * [new tag] viable/strict/1763635867 -> viable/strict/1763635867 2025-12-04T09:20:54.6422064Z * [new tag] viable/strict/1763639382 -> viable/strict/1763639382 2025-12-04T09:20:54.6422207Z * [new tag] viable/strict/1763646626 -> viable/strict/1763646626 2025-12-04T09:20:54.6422709Z * [new tag] viable/strict/1763655997 -> viable/strict/1763655997 2025-12-04T09:20:54.6425389Z * [new tag] viable/strict/1763659444 -> viable/strict/1763659444 2025-12-04T09:20:54.6425738Z * [new tag] viable/strict/1763660992 -> viable/strict/1763660992 2025-12-04T09:20:54.6425912Z * [new tag] viable/strict/1763663201 -> viable/strict/1763663201 2025-12-04T09:20:54.6426082Z * [new tag] viable/strict/1763670362 -> viable/strict/1763670362 2025-12-04T09:20:54.6426330Z * [new tag] viable/strict/1763675378 -> viable/strict/1763675378 2025-12-04T09:20:54.6432506Z * [new tag] viable/strict/1763693343 -> viable/strict/1763693343 2025-12-04T09:20:54.6437918Z * [new tag] viable/strict/1763696088 -> viable/strict/1763696088 2025-12-04T09:20:54.6438175Z * [new tag] viable/strict/1763697343 -> viable/strict/1763697343 2025-12-04T09:20:54.6438326Z * [new tag] viable/strict/1763699165 -> viable/strict/1763699165 2025-12-04T09:20:54.6438629Z * [new tag] viable/strict/1763700660 -> viable/strict/1763700660 2025-12-04T09:20:54.6438789Z * [new tag] viable/strict/1763704209 -> viable/strict/1763704209 2025-12-04T09:20:54.6439017Z * [new tag] viable/strict/1763706411 -> viable/strict/1763706411 2025-12-04T09:20:54.6439187Z * [new tag] viable/strict/1763708082 -> viable/strict/1763708082 2025-12-04T09:20:54.6439419Z * [new tag] viable/strict/1763711381 -> viable/strict/1763711381 2025-12-04T09:20:54.6439574Z * [new tag] viable/strict/1763713593 -> viable/strict/1763713593 2025-12-04T09:20:54.6439805Z * [new tag] viable/strict/1763715201 -> viable/strict/1763715201 2025-12-04T09:20:54.6439959Z * [new tag] viable/strict/1763733017 -> viable/strict/1763733017 2025-12-04T09:20:54.6440507Z * [new tag] viable/strict/1763735108 -> viable/strict/1763735108 2025-12-04T09:20:54.6440707Z * [new tag] viable/strict/1763749579 -> viable/strict/1763749579 2025-12-04T09:20:54.6440843Z * [new tag] viable/strict/1763751113 -> viable/strict/1763751113 2025-12-04T09:20:54.6440988Z * [new tag] viable/strict/1763753035 -> viable/strict/1763753035 2025-12-04T09:20:54.6441144Z * [new tag] viable/strict/1763754578 -> viable/strict/1763754578 2025-12-04T09:20:54.6441275Z * [new tag] viable/strict/1763756748 -> viable/strict/1763756748 2025-12-04T09:20:54.6441415Z * [new tag] viable/strict/1763758205 -> viable/strict/1763758205 2025-12-04T09:20:54.6441547Z * [new tag] viable/strict/1763764050 -> viable/strict/1763764050 2025-12-04T09:20:54.6441790Z * [new tag] viable/strict/1763771887 -> viable/strict/1763771887 2025-12-04T09:20:54.6442161Z * [new tag] viable/strict/1763773920 -> viable/strict/1763773920 2025-12-04T09:20:54.6442294Z * [new tag] viable/strict/1763776501 -> viable/strict/1763776501 2025-12-04T09:20:54.6442433Z * [new tag] viable/strict/1763779437 -> viable/strict/1763779437 2025-12-04T09:20:54.6442566Z * [new tag] viable/strict/1763781038 -> viable/strict/1763781038 2025-12-04T09:20:54.6442758Z * [new tag] viable/strict/1763782245 -> viable/strict/1763782245 2025-12-04T09:20:54.6442891Z * [new tag] viable/strict/1763785568 -> viable/strict/1763785568 2025-12-04T09:20:54.6443030Z * [new tag] viable/strict/1763787006 -> viable/strict/1763787006 2025-12-04T09:20:54.6448443Z * [new tag] viable/strict/1763789103 -> viable/strict/1763789103 2025-12-04T09:20:54.6448759Z * [new tag] viable/strict/1763790578 -> viable/strict/1763790578 2025-12-04T09:20:54.6448938Z * [new tag] viable/strict/1763796275 -> viable/strict/1763796275 2025-12-04T09:20:54.6449166Z * [new tag] viable/strict/1763801465 -> viable/strict/1763801465 2025-12-04T09:20:54.6449333Z * [new tag] viable/strict/1763803522 -> viable/strict/1763803522 2025-12-04T09:20:54.6449475Z * [new tag] viable/strict/1763808581 -> viable/strict/1763808581 2025-12-04T09:20:54.6454417Z * [new tag] viable/strict/1763840977 -> viable/strict/1763840977 2025-12-04T09:20:54.6460247Z * [new tag] viable/strict/1763846659 -> viable/strict/1763846659 2025-12-04T09:20:54.6464814Z * [new tag] viable/strict/1763872065 -> viable/strict/1763872065 2025-12-04T09:20:54.6469343Z * [new tag] viable/strict/1763873648 -> viable/strict/1763873648 2025-12-04T09:20:54.6473612Z * [new tag] viable/strict/1763875506 -> viable/strict/1763875506 2025-12-04T09:20:54.6478025Z * [new tag] viable/strict/1763889904 -> viable/strict/1763889904 2025-12-04T09:20:54.6479934Z * [new tag] viable/strict/1763930999 -> viable/strict/1763930999 2025-12-04T09:20:54.6480435Z * [new tag] viable/strict/1763944964 -> viable/strict/1763944964 2025-12-04T09:20:54.6480602Z * [new tag] viable/strict/1763958474 -> viable/strict/1763958474 2025-12-04T09:20:54.6480757Z * [new tag] viable/strict/1763967263 -> viable/strict/1763967263 2025-12-04T09:20:54.6480892Z * [new tag] viable/strict/1763972803 -> viable/strict/1763972803 2025-12-04T09:20:54.6481031Z * [new tag] viable/strict/1763976376 -> viable/strict/1763976376 2025-12-04T09:20:54.6481178Z * [new tag] viable/strict/1763989404 -> viable/strict/1763989404 2025-12-04T09:20:54.6481320Z * [new tag] viable/strict/1763990887 -> viable/strict/1763990887 2025-12-04T09:20:54.6481457Z * [new tag] viable/strict/1764019919 -> viable/strict/1764019919 2025-12-04T09:20:54.6481600Z * [new tag] viable/strict/1764023134 -> viable/strict/1764023134 2025-12-04T09:20:54.6481876Z * [new tag] viable/strict/1764024593 -> viable/strict/1764024593 2025-12-04T09:20:54.6482025Z * [new tag] viable/strict/1764026706 -> viable/strict/1764026706 2025-12-04T09:20:54.6482174Z * [new tag] viable/strict/1764031139 -> viable/strict/1764031139 2025-12-04T09:20:54.6482313Z * [new tag] viable/strict/1764033131 -> viable/strict/1764033131 2025-12-04T09:20:54.6482456Z * [new tag] viable/strict/1764035725 -> viable/strict/1764035725 2025-12-04T09:20:54.6482598Z * [new tag] viable/strict/1764624265 -> viable/strict/1764624265 2025-12-04T09:20:54.6482744Z * [new tag] viable/strict/1764631514 -> viable/strict/1764631514 2025-12-04T09:20:54.6483023Z * [new tag] viable/strict/1764632987 -> viable/strict/1764632987 2025-12-04T09:20:54.6483164Z * [new tag] viable/strict/1764636063 -> viable/strict/1764636063 2025-12-04T09:20:54.6483317Z * [new tag] viable/strict/1764643975 -> viable/strict/1764643975 2025-12-04T09:20:54.6483456Z * [new tag] viable/strict/1764646859 -> viable/strict/1764646859 2025-12-04T09:20:54.6483645Z * [new tag] viable/strict/1764653120 -> viable/strict/1764653120 2025-12-04T09:20:54.6483776Z * [new tag] viable/strict/1764654632 -> viable/strict/1764654632 2025-12-04T09:20:54.6483915Z * [new tag] viable/strict/1764656821 -> viable/strict/1764656821 2025-12-04T09:20:54.6484053Z * [new tag] viable/strict/1764658557 -> viable/strict/1764658557 2025-12-04T09:20:54.6484192Z * [new tag] viable/strict/1764660333 -> viable/strict/1764660333 2025-12-04T09:20:54.6484331Z * [new tag] viable/strict/1764661812 -> viable/strict/1764661812 2025-12-04T09:20:54.6484472Z * [new tag] viable/strict/1764664023 -> viable/strict/1764664023 2025-12-04T09:20:54.6484611Z * [new tag] viable/strict/1764669150 -> viable/strict/1764669150 2025-12-04T09:20:54.6484753Z * [new tag] viable/strict/1764680709 -> viable/strict/1764680709 2025-12-04T09:20:54.6484896Z * [new tag] viable/strict/1764687619 -> viable/strict/1764687619 2025-12-04T09:20:54.6485030Z * [new tag] viable/strict/1764696355 -> viable/strict/1764696355 2025-12-04T09:20:54.6485168Z * [new tag] viable/strict/1764701767 -> viable/strict/1764701767 2025-12-04T09:20:54.6485306Z * [new tag] viable/strict/1764710768 -> viable/strict/1764710768 2025-12-04T09:20:54.6485450Z * [new tag] viable/strict/1764716202 -> viable/strict/1764716202 2025-12-04T09:20:54.6485579Z * [new tag] viable/strict/1764793566 -> viable/strict/1764793566 2025-12-04T09:20:54.6485716Z * [new tag] viable/strict/1764797093 -> viable/strict/1764797093 2025-12-04T09:20:54.6485865Z * [new tag] viable/strict/1764800729 -> viable/strict/1764800729 2025-12-04T09:20:54.6485985Z * [new tag] whc_flight_1 -> whc_flight_1 2025-12-04T09:20:54.6486131Z * [new tag] whc_flight_2 -> whc_flight_2 2025-12-04T09:20:54.6486239Z * [new tag] whc_flight_4 -> whc_flight_4 2025-12-04T09:20:54.6997613Z [command]/usr/bin/git rev-parse --verify --quiet ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32^{object} 2025-12-04T09:20:54.7024879Z ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:20:54.7025877Z ##[endgroup] 2025-12-04T09:20:54.7026111Z ##[group]Determining the checkout info 2025-12-04T09:20:54.7028069Z ##[endgroup] 2025-12-04T09:20:54.7031673Z [command]/usr/bin/git sparse-checkout disable 2025-12-04T09:20:54.7064394Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-12-04T09:20:54.7097534Z ##[group]Checking out the ref 2025-12-04T09:20:54.7100588Z [command]/usr/bin/git checkout --progress --force ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:20:55.7331142Z Updating files: 95% (19274/20121) 2025-12-04T09:20:55.7447230Z Updating files: 96% (19317/20121) 2025-12-04T09:20:55.7568778Z Updating files: 97% (19518/20121) 2025-12-04T09:20:55.7790834Z Updating files: 98% (19719/20121) 2025-12-04T09:20:55.7919825Z Updating files: 99% (19920/20121) 2025-12-04T09:20:55.7920099Z Updating files: 100% (20121/20121) 2025-12-04T09:20:55.7920309Z Updating files: 100% (20121/20121), done. 2025-12-04T09:20:55.8141019Z Note: switching to 'ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32'. 2025-12-04T09:20:55.8143403Z 2025-12-04T09:20:55.8143726Z You are in 'detached HEAD' state. You can look around, make experimental 2025-12-04T09:20:55.8144084Z changes and commit them, and you can discard any commits you make in this 2025-12-04T09:20:55.8144725Z state without impacting any branches by switching back to a branch. 2025-12-04T09:20:55.8144907Z 2025-12-04T09:20:55.8145044Z If you want to create a new branch to retain commits you create, you may 2025-12-04T09:20:55.8145345Z do so (now or later) by using -c with the switch command. Example: 2025-12-04T09:20:55.8145520Z 2025-12-04T09:20:55.8145693Z git switch -c 2025-12-04T09:20:55.8145832Z 2025-12-04T09:20:55.8145911Z Or undo this operation with: 2025-12-04T09:20:55.8146030Z 2025-12-04T09:20:55.8146102Z git switch - 2025-12-04T09:20:55.8146193Z 2025-12-04T09:20:55.8146343Z Turn off this advice by setting config variable advice.detachedHead to false 2025-12-04T09:20:55.8146552Z 2025-12-04T09:20:55.8146714Z HEAD is now at ffd9b0fb435 Resolve collective autotuning test failure on arm (#168919) 2025-12-04T09:20:55.8202450Z ##[endgroup] 2025-12-04T09:20:55.8202817Z ##[group]Setting up auth for fetching submodules 2025-12-04T09:20:55.8214500Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-12-04T09:20:55.8285570Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-12-04T09:20:55.8310265Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-12-04T09:20:55.8342604Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-12-04T09:20:55.8367214Z ##[endgroup] 2025-12-04T09:20:55.8367515Z ##[group]Fetching submodules 2025-12-04T09:20:55.8373131Z [command]/usr/bin/git submodule sync --recursive 2025-12-04T09:20:55.8690083Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2025-12-04T09:20:55.9007054Z Submodule 'android/libs/fbjni' (https://github.com/facebookincubator/fbjni.git) registered for path 'android/libs/fbjni' 2025-12-04T09:20:55.9010407Z Submodule 'third_party/NNPACK_deps/FP16' (https://github.com/Maratyszcza/FP16.git) registered for path 'third_party/FP16' 2025-12-04T09:20:55.9011154Z Submodule 'third_party/NNPACK_deps/FXdiv' (https://github.com/Maratyszcza/FXdiv.git) registered for path 'third_party/FXdiv' 2025-12-04T09:20:55.9011887Z Submodule 'third_party/NNPACK' (https://github.com/Maratyszcza/NNPACK.git) registered for path 'third_party/NNPACK' 2025-12-04T09:20:55.9012470Z Submodule 'third_party/NVTX' (https://github.com/NVIDIA/NVTX.git) registered for path 'third_party/NVTX' 2025-12-04T09:20:55.9013239Z Submodule 'third_party/VulkanMemoryAllocator' (https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git) registered for path 'third_party/VulkanMemoryAllocator' 2025-12-04T09:20:55.9013999Z Submodule 'third_party/XNNPACK' (https://github.com/google/XNNPACK.git) registered for path 'third_party/XNNPACK' 2025-12-04T09:20:55.9014583Z Submodule 'third_party/aiter' (https://github.com/ROCm/aiter.git) registered for path 'third_party/aiter' 2025-12-04T09:20:55.9015247Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/benchmark' 2025-12-04T09:20:55.9016007Z Submodule 'third_party/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/composable_kernel' 2025-12-04T09:20:55.9017156Z Submodule 'third_party/cpp-httplib' (https://github.com/yhirose/cpp-httplib.git) registered for path 'third_party/cpp-httplib' 2025-12-04T09:20:55.9017894Z Submodule 'third_party/cpuinfo' (https://github.com/pytorch/cpuinfo.git) registered for path 'third_party/cpuinfo' 2025-12-04T09:20:55.9019602Z Submodule 'third_party/cudnn_frontend' (https://github.com/NVIDIA/cudnn-frontend.git) registered for path 'third_party/cudnn_frontend' 2025-12-04T09:20:55.9021695Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/cutlass' 2025-12-04T09:20:55.9024926Z Submodule 'third_party/fbgemm' (https://github.com/pytorch/fbgemm) registered for path 'third_party/fbgemm' 2025-12-04T09:20:55.9026174Z Submodule 'third_party/flash-attention' (https://github.com/Dao-AILab/flash-attention.git) registered for path 'third_party/flash-attention' 2025-12-04T09:20:55.9029720Z Submodule 'third_party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'third_party/flatbuffers' 2025-12-04T09:20:55.9033269Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/fmt' 2025-12-04T09:20:55.9034136Z Submodule 'third_party/gemmlowp/gemmlowp' (https://github.com/google/gemmlowp.git) registered for path 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:20:55.9036660Z Submodule 'third_party/gloo' (https://github.com/pytorch/gloo) registered for path 'third_party/gloo' 2025-12-04T09:20:55.9039106Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/googletest' 2025-12-04T09:20:55.9041275Z Submodule 'third_party/ideep' (https://github.com/intel/ideep) registered for path 'third_party/ideep' 2025-12-04T09:20:55.9046249Z Submodule 'third_party/ittapi' (https://github.com/intel/ittapi.git) registered for path 'third_party/ittapi' 2025-12-04T09:20:55.9046909Z Submodule 'third_party/kineto' (https://github.com/pytorch/kineto) registered for path 'third_party/kineto' 2025-12-04T09:20:55.9053315Z Submodule 'third_party/kleidiai' (https://github.com/ARM-software/kleidiai.git) registered for path 'third_party/kleidiai' 2025-12-04T09:20:55.9058254Z Submodule 'third_party/mimalloc' (https://github.com/microsoft/mimalloc.git) registered for path 'third_party/mimalloc' 2025-12-04T09:20:55.9059930Z Submodule 'third_party/nlohmann' (https://github.com/nlohmann/json.git) registered for path 'third_party/nlohmann' 2025-12-04T09:20:55.9060577Z Submodule 'third_party/onnx' (https://github.com/onnx/onnx.git) registered for path 'third_party/onnx' 2025-12-04T09:20:55.9064087Z Submodule 'third_party/opentelemetry-cpp' (https://github.com/open-telemetry/opentelemetry-cpp.git) registered for path 'third_party/opentelemetry-cpp' 2025-12-04T09:20:55.9064888Z Submodule 'third_party/pocketfft' (https://github.com/mreineck/pocketfft) registered for path 'third_party/pocketfft' 2025-12-04T09:20:55.9070706Z Submodule 'third_party/protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'third_party/protobuf' 2025-12-04T09:20:55.9072591Z Submodule 'third_party/NNPACK_deps/psimd' (https://github.com/Maratyszcza/psimd.git) registered for path 'third_party/psimd' 2025-12-04T09:20:55.9073386Z Submodule 'third_party/NNPACK_deps/pthreadpool' (https://github.com/Maratyszcza/pthreadpool.git) registered for path 'third_party/pthreadpool' 2025-12-04T09:20:55.9078397Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/pybind11' 2025-12-04T09:20:55.9079199Z Submodule 'third_party/python-peachpy' (https://github.com/malfet/PeachPy.git) registered for path 'third_party/python-peachpy' 2025-12-04T09:20:55.9080241Z Submodule 'third_party/sleef' (https://github.com/shibatch/sleef) registered for path 'third_party/sleef' 2025-12-04T09:20:55.9080835Z Submodule 'third_party/tensorpipe' (https://github.com/pytorch/tensorpipe.git) registered for path 'third_party/tensorpipe' 2025-12-04T09:20:55.9116305Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/android/libs/fbjni'... 2025-12-04T09:20:56.1123859Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FXdiv'... 2025-12-04T09:20:56.1124381Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FP16'... 2025-12-04T09:20:56.1124835Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/psimd'... 2025-12-04T09:20:56.1125271Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NNPACK'... 2025-12-04T09:20:56.1153443Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pybind11'... 2025-12-04T09:20:57.4779179Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pocketfft'... 2025-12-04T09:20:57.4780232Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep'... 2025-12-04T09:20:57.4781012Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/benchmark'... 2025-12-04T09:20:57.4781827Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pthreadpool'... 2025-12-04T09:20:57.4782786Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gemmlowp/gemmlowp'... 2025-12-04T09:20:57.4783623Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gloo'... 2025-12-04T09:20:57.4784378Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ittapi'... 2025-12-04T09:20:57.4785185Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/python-peachpy'... 2025-12-04T09:20:57.4786020Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kleidiai'... 2025-12-04T09:20:57.4786805Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NVTX'... 2025-12-04T09:20:57.4787572Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpuinfo'... 2025-12-04T09:20:57.4788383Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe'... 2025-12-04T09:20:57.4789225Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention'... 2025-12-04T09:20:57.4790077Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpp-httplib'... 2025-12-04T09:20:57.4790851Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/sleef'... 2025-12-04T09:20:57.4791624Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/mimalloc'... 2025-12-04T09:20:57.4792432Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/googletest'... 2025-12-04T09:20:57.5780039Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/VulkanMemoryAllocator'... 2025-12-04T09:20:58.1148842Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cudnn_frontend'... 2025-12-04T09:20:58.1149666Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto'... 2025-12-04T09:20:58.1150832Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fmt'... 2025-12-04T09:20:58.1151724Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flatbuffers'... 2025-12-04T09:20:58.2150875Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/XNNPACK'... 2025-12-04T09:21:11.8568716Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm'... 2025-12-04T09:21:11.8569310Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx'... 2025-12-04T09:21:11.8569872Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cutlass'... 2025-12-04T09:21:11.8570560Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/composable_kernel'... 2025-12-04T09:21:11.8571104Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp'... 2025-12-04T09:21:11.8571598Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/nlohmann'... 2025-12-04T09:21:11.8572105Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf'... 2025-12-04T09:21:11.9572014Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter'... 2025-12-04T09:21:12.2809993Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2025-12-04T09:21:12.2919722Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-12-04T09:21:12.3007593Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-12-04T09:21:12.3206540Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2025-12-04T09:21:12.3899604Z Submodule path 'third_party/NVTX': checked out '3ebbc93ded7285963bff932c678fa367eb393ba6' 2025-12-04T09:21:12.4342293Z Submodule path 'third_party/VulkanMemoryAllocator': checked out '1d8f600fd424278486eade7ed3e877c99f0846b1' 2025-12-04T09:21:12.9529295Z Submodule path 'third_party/XNNPACK': checked out '51a0103656eff6fc9bfd39a4597923c4b542c883' 2025-12-04T09:21:13.1066730Z Submodule path 'third_party/aiter': checked out '01aae101b9e5e94d6c16a9514c9fb8df99c93150' 2025-12-04T09:21:13.1080875Z Submodule '3rdparty/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:21:13.1114960Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter/3rdparty/composable_kernel'... 2025-12-04T09:21:18.1187894Z Submodule path 'third_party/aiter/3rdparty/composable_kernel': checked out 'cffe8fa2a442ac8e80dd236a1a5d24fe3d7e0cbf' 2025-12-04T09:21:18.1395417Z Submodule path 'third_party/benchmark': checked out '299e5928955cc62af9968370293b916f5130916f' 2025-12-04T09:21:18.4144626Z Submodule path 'third_party/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-12-04T09:21:18.4562988Z Submodule path 'third_party/cpp-httplib': checked out '89c932f313c6437c38f2982869beacc89c2f2246' 2025-12-04T09:21:18.5418753Z Submodule path 'third_party/cpuinfo': checked out 'f858c30bcb16f8effd5ff46996f0514539e17abc' 2025-12-04T09:21:18.5848402Z Submodule path 'third_party/cudnn_frontend': checked out '0b1577c8c83401237d601d0d0db5210506705396' 2025-12-04T09:21:19.1156168Z Submodule path 'third_party/cutlass': checked out 'f88806b1e31dfa579842638740216dd41fc6c588' 2025-12-04T09:21:19.2402408Z Submodule path 'third_party/fbgemm': checked out 'c0b988d39a9e47c794d699f29930ed4d7c7e13a4' 2025-12-04T09:21:19.2422148Z Submodule 'external/asmjit' (https://github.com/asmjit/asmjit.git) registered for path 'third_party/fbgemm/external/asmjit' 2025-12-04T09:21:19.2423018Z Submodule 'external/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:21:19.2428283Z Submodule 'external/cpuinfo' (https://github.com/pytorch/cpuinfo) registered for path 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:21:19.2431085Z Submodule 'external/cutlass' (https://github.com/jwfromm/cutlass) registered for path 'third_party/fbgemm/external/cutlass' 2025-12-04T09:21:19.2436056Z Submodule 'external/googletest' (https://github.com/google/googletest) registered for path 'third_party/fbgemm/external/googletest' 2025-12-04T09:21:19.2439749Z Submodule 'external/hipify_torch' (https://github.com/ROCmSoftwarePlatform/hipify_torch.git) registered for path 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:21:19.2440569Z Submodule 'external/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/fbgemm/external/json' 2025-12-04T09:21:19.2468250Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/asmjit'... 2025-12-04T09:21:20.3129292Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/hipify_torch'... 2025-12-04T09:21:20.3129923Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cpuinfo'... 2025-12-04T09:21:20.3130703Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/googletest'... 2025-12-04T09:21:20.4132555Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/composable_kernel'... 2025-12-04T09:21:24.0153692Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cutlass'... 2025-12-04T09:21:24.1153414Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/json'... 2025-12-04T09:21:26.3313317Z Submodule path 'third_party/fbgemm/external/asmjit': checked out 'a3199e8857792cd10b7589ff5d58343d2c9008ea' 2025-12-04T09:21:26.6103418Z Submodule path 'third_party/fbgemm/external/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-12-04T09:21:26.6979205Z Submodule path 'third_party/fbgemm/external/cpuinfo': checked out '6543fec09b2f04ac4a666882998b534afc9c1349' 2025-12-04T09:21:27.2025716Z Submodule path 'third_party/fbgemm/external/cutlass': checked out '98125ce499b0fdf7ffbe0e3052f5b8709f4840f8' 2025-12-04T09:21:27.2441332Z Submodule path 'third_party/fbgemm/external/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:21:27.2546816Z Submodule path 'third_party/fbgemm/external/hipify_torch': checked out '63b6a7b541fa7f08f8475ca7d74054db36ff2691' 2025-12-04T09:21:27.3394592Z Submodule path 'third_party/fbgemm/external/json': checked out '9cca280a4d0ccf0c08f47a99aa71d1b0e52f8d03' 2025-12-04T09:21:27.3962348Z Submodule path 'third_party/flash-attention': checked out '979702c87a8713a8e0a5e9fee122b90d2ef13be5' 2025-12-04T09:21:27.3978379Z Submodule 'csrc/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:21:27.3980699Z Submodule 'csrc/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:21:27.4011628Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/composable_kernel'... 2025-12-04T09:21:32.1487947Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/cutlass'... 2025-12-04T09:21:32.3236483Z Submodule path 'third_party/flash-attention/csrc/composable_kernel': checked out '888317e698e9803c62bd38568abc9e05d7709f33' 2025-12-04T09:21:32.7629616Z Submodule path 'third_party/flash-attention/csrc/cutlass': checked out 'c506e16788cb08416a4a57e11a9067beeee29420' 2025-12-04T09:21:32.8663416Z Submodule path 'third_party/flatbuffers': checked out 'a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757' 2025-12-04T09:21:32.8912760Z Submodule path 'third_party/fmt': checked out '407c905e45ad75fc29bf0f9bb7c5c2fd3475976f' 2025-12-04T09:21:32.9233140Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2025-12-04T09:21:32.9433190Z Submodule path 'third_party/gloo': checked out '54cbae0d3a67fa890b4c3d9ee162b7860315e341' 2025-12-04T09:21:32.9818376Z Submodule path 'third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:21:32.9928434Z Submodule path 'third_party/ideep': checked out '719d8e6cd7f7a0e01b155657526d693acf97c2b3' 2025-12-04T09:21:32.9944036Z Submodule 'mkl-dnn' (https://github.com/intel/mkl-dnn.git) registered for path 'third_party/ideep/mkl-dnn' 2025-12-04T09:21:32.9967564Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep/mkl-dnn'... 2025-12-04T09:21:49.1000806Z Submodule path 'third_party/ideep/mkl-dnn': checked out '8d263e693366ef8db40acc569cc7d8edf644556d' 2025-12-04T09:21:49.1170733Z Submodule path 'third_party/ittapi': checked out 'dec1d23ca65ab069d225dfe40dea14f455170959' 2025-12-04T09:21:49.1976418Z Submodule path 'third_party/kineto': checked out '31f85df8fbd89c188f14ef10f1ec65379786b943' 2025-12-04T09:21:49.1994110Z Submodule 'libkineto/third_party/dynolog' (https://github.com/facebookincubator/dynolog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:21:49.1996040Z Submodule 'libkineto/third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:21:49.1996903Z Submodule 'libkineto/third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:21:49.2025117Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog'... 2025-12-04T09:21:50.5519623Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/fmt'... 2025-12-04T09:21:50.6696929Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/googletest'... 2025-12-04T09:21:50.7431686Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out 'd2ffe0a4e3acace628db49974246b66fc3e85fb1' 2025-12-04T09:21:50.7447665Z Submodule 'third_party/DCGM' (https://github.com/NVIDIA/DCGM.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:21:50.7448586Z Submodule 'third_party/cpr' (https://github.com/libcpr/cpr.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:21:50.7449220Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:21:50.7449872Z Submodule 'third_party/gflags' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:21:50.7454872Z Submodule 'third_party/glog' (https://github.com/google/glog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:21:50.7456763Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:21:50.7459205Z Submodule 'third_party/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:21:50.7460040Z Submodule 'third_party/pfs' (https://github.com/dtrugman/pfs.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:21:50.7465448Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:21:50.7485292Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM'... 2025-12-04T09:21:52.3857437Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/pfs'... 2025-12-04T09:21:52.3858648Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp'... 2025-12-04T09:21:52.3859378Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags'... 2025-12-04T09:21:52.3859999Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/cpr'... 2025-12-04T09:21:52.3860590Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/glog'... 2025-12-04T09:21:52.3861198Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/googletest'... 2025-12-04T09:21:52.3861807Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/fmt'... 2025-12-04T09:21:52.4858737Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/json'... 2025-12-04T09:21:57.3116593Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2025-12-04T09:21:57.3265871Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2025-12-04T09:21:57.3569175Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2025-12-04T09:21:57.3685896Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2025-12-04T09:21:57.3700409Z Submodule 'doc' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:21:57.3732134Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc'... 2025-12-04T09:21:57.9760440Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2025-12-04T09:21:57.9926192Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2025-12-04T09:21:58.0294206Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:21:58.1106919Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2025-12-04T09:21:58.1246168Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2025-12-04T09:21:58.1392472Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp': checked out 'b1234816facfdda29845c46696a02998a4af115a' 2025-12-04T09:21:58.1405914Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:21:58.1406866Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:21:58.1437195Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-12-04T09:21:59.9923818Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest'... 2025-12-04T09:22:00.1998962Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'd7ba35bbb649209c66e582d5a0244ba988a15159' 2025-12-04T09:22:00.2374858Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-12-04T09:22:00.2653993Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out '40626af88bd7df9a5fb80be7b25ac85b122d6c21' 2025-12-04T09:22:00.3025985Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:22:00.3464497Z Submodule path 'third_party/kleidiai': checked out 'd7770c89632329a9914ef1a90289917597639cbe' 2025-12-04T09:22:00.3777128Z Submodule path 'third_party/mimalloc': checked out 'fbd8b99c2b828428947d70fdc046bb55609be93e' 2025-12-04T09:22:00.4599952Z Submodule path 'third_party/nlohmann': checked out '55f93686c01528224f448c19128836e7df245f72' 2025-12-04T09:22:00.7277516Z Submodule path 'third_party/onnx': checked out 'e709452ef2bbc1d113faf678c24e6d3467696e83' 2025-12-04T09:22:00.7310217Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/onnx/third_party/pybind11' 2025-12-04T09:22:00.7334439Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx/third_party/pybind11'... 2025-12-04T09:22:01.5577354Z Submodule path 'third_party/onnx/third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-12-04T09:22:01.6097857Z Submodule path 'third_party/opentelemetry-cpp': checked out 'a799f4aed9c94b765dcdaabaeab7d5e7e2310878' 2025-12-04T09:22:01.6114484Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark) registered for path 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:22:01.6117451Z Submodule 'third_party/googletest' (https://github.com/google/googletest) registered for path 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:22:01.6118531Z Submodule 'third_party/ms-gsl' (https://github.com/microsoft/GSL) registered for path 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:22:01.6119315Z Submodule 'third_party/nlohmann-json' (https://github.com/nlohmann/json) registered for path 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:22:01.6120083Z Submodule 'third_party/opentelemetry-proto' (https://github.com/open-telemetry/opentelemetry-proto) registered for path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:22:01.6120895Z Submodule 'third_party/opentracing-cpp' (https://github.com/opentracing/opentracing-cpp.git) registered for path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:22:01.6121791Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:22:01.6122460Z Submodule 'tools/vcpkg' (https://github.com/Microsoft/vcpkg) registered for path 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:22:01.6152062Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/benchmark'... 2025-12-04T09:22:01.9931473Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentracing-cpp'... 2025-12-04T09:22:01.9932240Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentelemetry-proto'... 2025-12-04T09:22:01.9932877Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/ms-gsl'... 2025-12-04T09:22:01.9933528Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp'... 2025-12-04T09:22:02.0934175Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/googletest'... 2025-12-04T09:22:02.6612638Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/nlohmann-json'... 2025-12-04T09:22:08.2963789Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/tools/vcpkg'... 2025-12-04T09:22:09.1683842Z Submodule path 'third_party/opentelemetry-cpp/third_party/benchmark': checked out 'd572f4777349d43653b21d6c2fc63020ab326db2' 2025-12-04T09:22:09.2025794Z Submodule path 'third_party/opentelemetry-cpp/third_party/googletest': checked out 'b796f7d44681514f58a683a3a71ff17c94edb0c1' 2025-12-04T09:22:09.2179517Z Submodule path 'third_party/opentelemetry-cpp/third_party/ms-gsl': checked out '6f4529395c5b7c2d661812257cd6780c67e54afa' 2025-12-04T09:22:09.3020775Z Submodule path 'third_party/opentelemetry-cpp/third_party/nlohmann-json': checked out 'bc889afb4c5bf1c0d8ee29ef35eaaf4c8bef8a5d' 2025-12-04T09:22:09.3140502Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto': checked out '4ca4f0335c63cda7ab31ea7ed70d6553aee14dce' 2025-12-04T09:22:09.3267465Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp': checked out '06b57f48ded1fa3bdd3d4346f6ef29e40e08eaf5' 2025-12-04T09:22:09.3393347Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp': checked out 'c9ffcdda9086ffd9e1283ea7a0276d831f3c8a8d' 2025-12-04T09:22:09.3410738Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:09.3411572Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:09.3438149Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-12-04T09:22:11.3532623Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest'... 2025-12-04T09:22:11.5601076Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'eefb26f82b233268fc98577d265352720d477ba4' 2025-12-04T09:22:11.5977698Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-12-04T09:22:11.9021460Z Submodule path 'third_party/opentelemetry-cpp/tools/vcpkg': checked out '8eb57355a4ffb410a2e94c07b4dca2dffbee8e50' 2025-12-04T09:22:11.9130526Z Submodule path 'third_party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-12-04T09:22:12.1190491Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2025-12-04T09:22:12.1209271Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:22:12.1211384Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/protobuf/third_party/googletest' 2025-12-04T09:22:12.1236130Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/benchmark'... 2025-12-04T09:22:12.6704531Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/googletest'... 2025-12-04T09:22:13.0144095Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2025-12-04T09:22:13.0738110Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2025-12-04T09:22:13.0818337Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2025-12-04T09:22:13.0926178Z Submodule path 'third_party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-12-04T09:22:13.1265599Z Submodule path 'third_party/pybind11': checked out 'f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8' 2025-12-04T09:22:13.1495221Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2025-12-04T09:22:13.1845906Z Submodule path 'third_party/sleef': checked out '5a1d179df9cf652951b59010a2d2075372d67f68' 2025-12-04T09:22:13.2056691Z Submodule path 'third_party/tensorpipe': checked out '2b4cd91092d335a697416b2a3cb398283246849d' 2025-12-04T09:22:13.2070333Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:22:13.2075015Z Submodule 'third_party/libnop' (https://github.com/google/libnop.git) registered for path 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:22:13.2079996Z Submodule 'third_party/libuv' (https://github.com/libuv/libuv.git) registered for path 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:22:13.2080614Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:22:13.2101658Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/googletest'... 2025-12-04T09:22:14.1156274Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libnop'... 2025-12-04T09:22:14.1158500Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11'... 2025-12-04T09:22:14.1677830Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libuv'... 2025-12-04T09:22:14.2146939Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2025-12-04T09:22:14.2281168Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2025-12-04T09:22:14.2880797Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '5152db2cbfeb5582e9c27c5ea1dba2cd9e10759b' 2025-12-04T09:22:14.3118406Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2025-12-04T09:22:14.3129550Z Submodule 'tools/clang' (https://github.com/wjakob/clang-cindex-python3) registered for path 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:22:14.3154169Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11/tools/clang'... 2025-12-04T09:22:14.4883346Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2025-12-04T09:22:14.4924695Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-12-04T09:22:14.5221151Z Entering 'android/libs/fbjni' 2025-12-04T09:22:14.5265156Z Entering 'third_party/FP16' 2025-12-04T09:22:14.5308197Z Entering 'third_party/FXdiv' 2025-12-04T09:22:14.5346760Z Entering 'third_party/NNPACK' 2025-12-04T09:22:14.5386848Z Entering 'third_party/NVTX' 2025-12-04T09:22:14.5426684Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:22:14.5470763Z Entering 'third_party/XNNPACK' 2025-12-04T09:22:14.5520888Z Entering 'third_party/aiter' 2025-12-04T09:22:14.5563427Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:22:14.5607116Z Entering 'third_party/benchmark' 2025-12-04T09:22:14.5649591Z Entering 'third_party/composable_kernel' 2025-12-04T09:22:14.5695966Z Entering 'third_party/cpp-httplib' 2025-12-04T09:22:14.5736058Z Entering 'third_party/cpuinfo' 2025-12-04T09:22:14.5780863Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:22:14.5817700Z Entering 'third_party/cutlass' 2025-12-04T09:22:14.5863967Z Entering 'third_party/fbgemm' 2025-12-04T09:22:14.5905815Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:22:14.5942989Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:22:14.5990707Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:22:14.6027045Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:22:14.6068818Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:22:14.6108130Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:22:14.6146375Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:22:14.6190094Z Entering 'third_party/flash-attention' 2025-12-04T09:22:14.6231098Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:22:14.6277946Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:22:14.6325944Z Entering 'third_party/flatbuffers' 2025-12-04T09:22:14.6370676Z Entering 'third_party/fmt' 2025-12-04T09:22:14.6407522Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:22:14.6444027Z Entering 'third_party/gloo' 2025-12-04T09:22:14.6485963Z Entering 'third_party/googletest' 2025-12-04T09:22:14.6521017Z Entering 'third_party/ideep' 2025-12-04T09:22:14.6560927Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:22:14.6609930Z Entering 'third_party/ittapi' 2025-12-04T09:22:14.6650533Z Entering 'third_party/kineto' 2025-12-04T09:22:14.6686847Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:22:14.6722663Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:22:14.6759521Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:22:14.6802396Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:22:14.6845783Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:22:14.6883479Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:22:14.6926192Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:22:14.6968679Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:22:14.7003867Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:22:14.7040017Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:22:14.7079400Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:22:14.7119343Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:14.7161969Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:14.7207674Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:22:14.7247956Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:22:14.7290732Z Entering 'third_party/kleidiai' 2025-12-04T09:22:14.7325847Z Entering 'third_party/mimalloc' 2025-12-04T09:22:14.7365268Z Entering 'third_party/nlohmann' 2025-12-04T09:22:14.7404375Z Entering 'third_party/onnx' 2025-12-04T09:22:14.7457356Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:22:14.7498194Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:22:14.7539559Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:22:14.7576028Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:22:14.7613989Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:22:14.7650066Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:22:14.7689472Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:22:14.7725465Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:22:14.7760830Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:22:14.7793850Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:14.7830353Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:14.7874919Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:22:14.7932413Z Entering 'third_party/pocketfft' 2025-12-04T09:22:14.7972670Z Entering 'third_party/protobuf' 2025-12-04T09:22:14.8014254Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:22:14.8054880Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:22:14.8096162Z Entering 'third_party/psimd' 2025-12-04T09:22:14.8136513Z Entering 'third_party/pthreadpool' 2025-12-04T09:22:14.8178489Z Entering 'third_party/pybind11' 2025-12-04T09:22:14.8216994Z Entering 'third_party/python-peachpy' 2025-12-04T09:22:14.8258388Z Entering 'third_party/sleef' 2025-12-04T09:22:14.8299995Z Entering 'third_party/tensorpipe' 2025-12-04T09:22:14.8335434Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:22:14.8374522Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:22:14.8410416Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:22:14.8454379Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:22:14.8489447Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:22:14.8552615Z ##[endgroup] 2025-12-04T09:22:14.8553253Z ##[group]Persisting credentials for submodules 2025-12-04T09:22:14.8557403Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-12-04T09:22:14.8853704Z Entering 'android/libs/fbjni' 2025-12-04T09:22:14.8906780Z Entering 'third_party/FP16' 2025-12-04T09:22:14.8958840Z Entering 'third_party/FXdiv' 2025-12-04T09:22:14.9011697Z Entering 'third_party/NNPACK' 2025-12-04T09:22:14.9059644Z Entering 'third_party/NVTX' 2025-12-04T09:22:14.9112692Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:22:14.9168212Z Entering 'third_party/XNNPACK' 2025-12-04T09:22:14.9228347Z Entering 'third_party/aiter' 2025-12-04T09:22:14.9280947Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:22:14.9344191Z Entering 'third_party/benchmark' 2025-12-04T09:22:14.9396621Z Entering 'third_party/composable_kernel' 2025-12-04T09:22:14.9452848Z Entering 'third_party/cpp-httplib' 2025-12-04T09:22:14.9506680Z Entering 'third_party/cpuinfo' 2025-12-04T09:22:14.9559480Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:22:14.9607299Z Entering 'third_party/cutlass' 2025-12-04T09:22:14.9664362Z Entering 'third_party/fbgemm' 2025-12-04T09:22:14.9719170Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:22:14.9772553Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:22:14.9829118Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:22:14.9889118Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:22:14.9951108Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:22:15.0006096Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:22:15.0059664Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:22:15.0116585Z Entering 'third_party/flash-attention' 2025-12-04T09:22:15.0165351Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:22:15.0223112Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:22:15.0281802Z Entering 'third_party/flatbuffers' 2025-12-04T09:22:15.0340083Z Entering 'third_party/fmt' 2025-12-04T09:22:15.0392971Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:22:15.0441907Z Entering 'third_party/gloo' 2025-12-04T09:22:15.0499199Z Entering 'third_party/googletest' 2025-12-04T09:22:15.0548973Z Entering 'third_party/ideep' 2025-12-04T09:22:15.0599869Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:22:15.0657296Z Entering 'third_party/ittapi' 2025-12-04T09:22:15.0710084Z Entering 'third_party/kineto' 2025-12-04T09:22:15.0760719Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:22:15.0811170Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:22:15.0864974Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:22:15.0916140Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:22:15.0966407Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:22:15.1013864Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:22:15.1069264Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:22:15.1119365Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:22:15.1178206Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:22:15.1231123Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:22:15.1282338Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:22:15.1331672Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:15.1387952Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:15.1447294Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:22:15.1501712Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:22:15.1557264Z Entering 'third_party/kleidiai' 2025-12-04T09:22:15.1607649Z Entering 'third_party/mimalloc' 2025-12-04T09:22:15.1662989Z Entering 'third_party/nlohmann' 2025-12-04T09:22:15.1716357Z Entering 'third_party/onnx' 2025-12-04T09:22:15.1781733Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:22:15.1841632Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:22:15.1891385Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:22:15.1947320Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:22:15.2000745Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:22:15.2051758Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:22:15.2102743Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:22:15.2157414Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:22:15.2208421Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:22:15.2263032Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:15.2317039Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:15.2375285Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:22:15.2439924Z Entering 'third_party/pocketfft' 2025-12-04T09:22:15.2496184Z Entering 'third_party/protobuf' 2025-12-04T09:22:15.2548801Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:22:15.2599875Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:22:15.2652103Z Entering 'third_party/psimd' 2025-12-04T09:22:15.2707792Z Entering 'third_party/pthreadpool' 2025-12-04T09:22:15.2761674Z Entering 'third_party/pybind11' 2025-12-04T09:22:15.2813492Z Entering 'third_party/python-peachpy' 2025-12-04T09:22:15.2872654Z Entering 'third_party/sleef' 2025-12-04T09:22:15.2924041Z Entering 'third_party/tensorpipe' 2025-12-04T09:22:15.2976855Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:22:15.3022121Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:22:15.3079282Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:22:15.3136315Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:22:15.3183942Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:22:15.3256515Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-12-04T09:22:15.3562463Z Entering 'android/libs/fbjni' 2025-12-04T09:22:15.3611493Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T09:22:15.3626408Z Entering 'third_party/FP16' 2025-12-04T09:22:15.3671351Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T09:22:15.3688995Z Entering 'third_party/FXdiv' 2025-12-04T09:22:15.3737100Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T09:22:15.3749573Z Entering 'third_party/NNPACK' 2025-12-04T09:22:15.3798342Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T09:22:15.3815012Z Entering 'third_party/NVTX' 2025-12-04T09:22:15.3864228Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T09:22:15.3882018Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:22:15.3930574Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T09:22:15.3948480Z Entering 'third_party/XNNPACK' 2025-12-04T09:22:15.3993808Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T09:22:15.4020825Z Entering 'third_party/aiter' 2025-12-04T09:22:15.4068666Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T09:22:15.4082547Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:22:15.4125986Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T09:22:15.4156792Z Entering 'third_party/benchmark' 2025-12-04T09:22:15.4201183Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:22:15.4218278Z Entering 'third_party/composable_kernel' 2025-12-04T09:22:15.4267383Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T09:22:15.4288575Z Entering 'third_party/cpp-httplib' 2025-12-04T09:22:15.4334243Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T09:22:15.4353876Z Entering 'third_party/cpuinfo' 2025-12-04T09:22:15.4397528Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T09:22:15.4416723Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:22:15.4460999Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T09:22:15.4475123Z Entering 'third_party/cutlass' 2025-12-04T09:22:15.4518852Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T09:22:15.4544925Z Entering 'third_party/fbgemm' 2025-12-04T09:22:15.4590979Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T09:22:15.4604945Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:22:15.4649728Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T09:22:15.4667389Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:22:15.4713449Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T09:22:15.4741146Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:22:15.4789984Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T09:22:15.4806284Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:22:15.4851983Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T09:22:15.4877837Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:22:15.4921441Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T09:22:15.4944009Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:22:15.4987743Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T09:22:15.5003888Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:22:15.5052674Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T09:22:15.5073237Z Entering 'third_party/flash-attention' 2025-12-04T09:22:15.5118744Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T09:22:15.5138725Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:22:15.5180890Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T09:22:15.5204455Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:22:15.5248043Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T09:22:15.5272031Z Entering 'third_party/flatbuffers' 2025-12-04T09:22:15.5317167Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T09:22:15.5339926Z Entering 'third_party/fmt' 2025-12-04T09:22:15.5383874Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T09:22:15.5401900Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:22:15.5445766Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T09:22:15.5469063Z Entering 'third_party/gloo' 2025-12-04T09:22:15.5511463Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T09:22:15.5526278Z Entering 'third_party/googletest' 2025-12-04T09:22:15.5575275Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:22:15.5594345Z Entering 'third_party/ideep' 2025-12-04T09:22:15.5639463Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T09:22:15.5653473Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:22:15.5699567Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T09:22:15.5724020Z Entering 'third_party/ittapi' 2025-12-04T09:22:15.5772225Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T09:22:15.5791869Z Entering 'third_party/kineto' 2025-12-04T09:22:15.5838459Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T09:22:15.5857245Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:22:15.5897614Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T09:22:15.5912641Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:22:15.5958356Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T09:22:15.5977332Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:22:15.6026808Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T09:22:15.6042119Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:22:15.6089914Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T09:22:15.6108414Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:22:15.6156871Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T09:22:15.6172645Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:22:15.6224025Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T09:22:15.6243447Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:22:15.6288953Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T09:22:15.6310746Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:22:15.6357379Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:22:15.6379117Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:22:15.6415730Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T09:22:15.6441282Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:22:15.6485600Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T09:22:15.6503161Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:22:15.6549445Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T09:22:15.6565984Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:15.6610687Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T09:22:15.6630951Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:15.6680699Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T09:22:15.6703157Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:22:15.6750148Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T09:22:15.6763362Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:22:15.6807270Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T09:22:15.6827254Z Entering 'third_party/kleidiai' 2025-12-04T09:22:15.6872556Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T09:22:15.6892891Z Entering 'third_party/mimalloc' 2025-12-04T09:22:15.6934532Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T09:22:15.6953815Z Entering 'third_party/nlohmann' 2025-12-04T09:22:15.6999025Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T09:22:15.7019394Z Entering 'third_party/onnx' 2025-12-04T09:22:15.7065923Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T09:22:15.7092381Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:22:15.7137812Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:22:15.7156052Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:22:15.7203712Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T09:22:15.7221134Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:22:15.7267788Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:22:15.7283792Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:22:15.7330499Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:22:15.7347034Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:22:15.7397998Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T09:22:15.7413692Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:22:15.7459653Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T09:22:15.7477721Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:22:15.7523843Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T09:22:15.7545225Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:22:15.7588211Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T09:22:15.7603074Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:22:15.7651326Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T09:22:15.7666327Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:15.7709061Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T09:22:15.7727380Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:15.7773266Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T09:22:15.7793241Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:22:15.7840640Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T09:22:15.7876079Z Entering 'third_party/pocketfft' 2025-12-04T09:22:15.7919007Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T09:22:15.7935209Z Entering 'third_party/protobuf' 2025-12-04T09:22:15.7980622Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T09:22:15.7998529Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:22:15.8048548Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:22:15.8063191Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:22:15.8108678Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:22:15.8134965Z Entering 'third_party/psimd' 2025-12-04T09:22:15.8174701Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T09:22:15.8191472Z Entering 'third_party/pthreadpool' 2025-12-04T09:22:15.8236435Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T09:22:15.8254485Z Entering 'third_party/pybind11' 2025-12-04T09:22:15.8301782Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:22:15.8317824Z Entering 'third_party/python-peachpy' 2025-12-04T09:22:15.8368199Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T09:22:15.8385644Z Entering 'third_party/sleef' 2025-12-04T09:22:15.8431818Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T09:22:15.8452710Z Entering 'third_party/tensorpipe' 2025-12-04T09:22:15.8499528Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T09:22:15.8512728Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:22:15.8557978Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:22:15.8575327Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:22:15.8619399Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T09:22:15.8635942Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:22:15.8686444Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T09:22:15.8700556Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:22:15.8747514Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:22:15.8762042Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:22:15.8807623Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T09:22:16.0709627Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-12-04T09:22:16.1015987Z Entering 'android/libs/fbjni' 2025-12-04T09:22:16.1055138Z Entering 'third_party/FP16' 2025-12-04T09:22:16.1094014Z Entering 'third_party/FXdiv' 2025-12-04T09:22:16.1131548Z Entering 'third_party/NNPACK' 2025-12-04T09:22:16.1186037Z Entering 'third_party/NVTX' 2025-12-04T09:22:16.1225240Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:22:16.1263463Z Entering 'third_party/XNNPACK' 2025-12-04T09:22:16.1313153Z Entering 'third_party/aiter' 2025-12-04T09:22:16.1347793Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:22:16.1394887Z Entering 'third_party/benchmark' 2025-12-04T09:22:16.1439468Z Entering 'third_party/composable_kernel' 2025-12-04T09:22:16.1485789Z Entering 'third_party/cpp-httplib' 2025-12-04T09:22:16.1522985Z Entering 'third_party/cpuinfo' 2025-12-04T09:22:16.1563989Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:22:16.1605049Z Entering 'third_party/cutlass' 2025-12-04T09:22:16.1653746Z Entering 'third_party/fbgemm' 2025-12-04T09:22:16.1695079Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:22:16.1730904Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:22:16.1779091Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:22:16.1818671Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:22:16.1863483Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:22:16.1897303Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:22:16.1934518Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:22:16.1976094Z Entering 'third_party/flash-attention' 2025-12-04T09:22:16.2013555Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:22:16.2051937Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:22:16.2101463Z Entering 'third_party/flatbuffers' 2025-12-04T09:22:16.2144869Z Entering 'third_party/fmt' 2025-12-04T09:22:16.2182118Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:22:16.2224597Z Entering 'third_party/gloo' 2025-12-04T09:22:16.2268306Z Entering 'third_party/googletest' 2025-12-04T09:22:16.2309084Z Entering 'third_party/ideep' 2025-12-04T09:22:16.2344154Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:22:16.2386160Z Entering 'third_party/ittapi' 2025-12-04T09:22:16.2427613Z Entering 'third_party/kineto' 2025-12-04T09:22:16.2466359Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:22:16.2502089Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:22:16.2537432Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:22:16.2577900Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:22:16.2623722Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:22:16.2660600Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:22:16.2705200Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:22:16.2747554Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:22:16.2782358Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:22:16.2822567Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:22:16.2858413Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:22:16.2901372Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:16.2935247Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:16.2981208Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:22:16.3023063Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:22:16.3058468Z Entering 'third_party/kleidiai' 2025-12-04T09:22:16.3103212Z Entering 'third_party/mimalloc' 2025-12-04T09:22:16.3143264Z Entering 'third_party/nlohmann' 2025-12-04T09:22:16.3180523Z Entering 'third_party/onnx' 2025-12-04T09:22:16.3231435Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:22:16.3275885Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:22:16.3315738Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:22:16.3356272Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:22:16.3390932Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:22:16.3427418Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:22:16.3470409Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:22:16.3505152Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:22:16.3544991Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:22:16.3582862Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:16.3619683Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:16.3662873Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:22:16.3719755Z Entering 'third_party/pocketfft' 2025-12-04T09:22:16.3759421Z Entering 'third_party/protobuf' 2025-12-04T09:22:16.3798731Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:22:16.3838679Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:22:16.3882004Z Entering 'third_party/psimd' 2025-12-04T09:22:16.3922071Z Entering 'third_party/pthreadpool' 2025-12-04T09:22:16.3963840Z Entering 'third_party/pybind11' 2025-12-04T09:22:16.4002872Z Entering 'third_party/python-peachpy' 2025-12-04T09:22:16.4040393Z Entering 'third_party/sleef' 2025-12-04T09:22:16.4082232Z Entering 'third_party/tensorpipe' 2025-12-04T09:22:16.4116999Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:22:16.4154899Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:22:16.4192697Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:22:16.4231450Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:22:16.4265155Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:22:16.4322461Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-12-04T09:22:16.4624371Z Entering 'android/libs/fbjni' 2025-12-04T09:22:16.4666441Z Entering 'third_party/FP16' 2025-12-04T09:22:16.4702611Z Entering 'third_party/FXdiv' 2025-12-04T09:22:16.4740619Z Entering 'third_party/NNPACK' 2025-12-04T09:22:16.4786162Z Entering 'third_party/NVTX' 2025-12-04T09:22:16.4826504Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:22:16.4868161Z Entering 'third_party/XNNPACK' 2025-12-04T09:22:16.4917329Z Entering 'third_party/aiter' 2025-12-04T09:22:16.4953674Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:22:16.5009000Z Entering 'third_party/benchmark' 2025-12-04T09:22:16.5047784Z Entering 'third_party/composable_kernel' 2025-12-04T09:22:16.5094907Z Entering 'third_party/cpp-httplib' 2025-12-04T09:22:16.5138992Z Entering 'third_party/cpuinfo' 2025-12-04T09:22:16.5179745Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:22:16.5219390Z Entering 'third_party/cutlass' 2025-12-04T09:22:16.5268022Z Entering 'third_party/fbgemm' 2025-12-04T09:22:16.5310084Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:22:16.5344441Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:22:16.5391360Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:22:16.5425904Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:22:16.5470511Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:22:16.5515723Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:22:16.5552587Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:22:16.5598469Z Entering 'third_party/flash-attention' 2025-12-04T09:22:16.5635512Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:22:16.5676764Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:22:16.5726317Z Entering 'third_party/flatbuffers' 2025-12-04T09:22:16.5768168Z Entering 'third_party/fmt' 2025-12-04T09:22:16.5805781Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:22:16.5843315Z Entering 'third_party/gloo' 2025-12-04T09:22:16.5889256Z Entering 'third_party/googletest' 2025-12-04T09:22:16.5928534Z Entering 'third_party/ideep' 2025-12-04T09:22:16.5966325Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:22:16.6016582Z Entering 'third_party/ittapi' 2025-12-04T09:22:16.6056539Z Entering 'third_party/kineto' 2025-12-04T09:22:16.6094299Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:22:16.6131317Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:22:16.6172650Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:22:16.6209404Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:22:16.6248765Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:22:16.6288140Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:22:16.6325786Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:22:16.6369083Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:22:16.6410442Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:22:16.6448767Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:22:16.6482887Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:22:16.6521489Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:16.6563564Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:16.6608208Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:22:16.6644480Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:22:16.6686820Z Entering 'third_party/kleidiai' 2025-12-04T09:22:16.6724590Z Entering 'third_party/mimalloc' 2025-12-04T09:22:16.6768169Z Entering 'third_party/nlohmann' 2025-12-04T09:22:16.6808447Z Entering 'third_party/onnx' 2025-12-04T09:22:16.6859798Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:22:16.6897845Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:22:16.6939940Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:22:16.6977029Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:22:16.7014529Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:22:16.7052147Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:22:16.7087872Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:22:16.7125034Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:22:16.7167249Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:22:16.7199545Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:16.7241422Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:16.7284815Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:22:16.7343352Z Entering 'third_party/pocketfft' 2025-12-04T09:22:16.7385512Z Entering 'third_party/protobuf' 2025-12-04T09:22:16.7427677Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:22:16.7468424Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:22:16.7510331Z Entering 'third_party/psimd' 2025-12-04T09:22:16.7549045Z Entering 'third_party/pthreadpool' 2025-12-04T09:22:16.7589335Z Entering 'third_party/pybind11' 2025-12-04T09:22:16.7631248Z Entering 'third_party/python-peachpy' 2025-12-04T09:22:16.7671890Z Entering 'third_party/sleef' 2025-12-04T09:22:16.7709839Z Entering 'third_party/tensorpipe' 2025-12-04T09:22:16.7746268Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:22:16.7782304Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:22:16.7822386Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:22:16.7863514Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:22:16.7898067Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:22:16.7959756Z ##[endgroup] 2025-12-04T09:22:16.7997345Z [command]/usr/bin/git log -1 --format=%H 2025-12-04T09:22:16.8015475Z ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:22:16.8133153Z ##[group]Run cd "${GITHUB_WORKSPACE}" 2025-12-04T09:22:16.8133403Z cd "${GITHUB_WORKSPACE}" 2025-12-04T09:22:16.8133600Z # Clean stale submodule dirs 2025-12-04T09:22:16.8133805Z if [ -z "${NO_SUDO}" ]; then 2025-12-04T09:22:16.8134045Z  sudo git submodule foreach --recursive git clean -ffdx 2025-12-04T09:22:16.8134272Z else 2025-12-04T09:22:16.8134484Z  git submodule foreach --recursive git clean -ffdx 2025-12-04T09:22:16.8134708Z fi 2025-12-04T09:22:16.8141552Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:16.8141790Z env: 2025-12-04T09:22:16.8141943Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:16.8142106Z NO_SUDO: true 2025-12-04T09:22:16.8142261Z ##[endgroup] 2025-12-04T09:22:16.8464104Z Entering 'android/libs/fbjni' 2025-12-04T09:22:16.8498220Z Entering 'third_party/FP16' 2025-12-04T09:22:16.8531203Z Entering 'third_party/FXdiv' 2025-12-04T09:22:16.8558912Z Entering 'third_party/NNPACK' 2025-12-04T09:22:16.8592825Z Entering 'third_party/NVTX' 2025-12-04T09:22:16.8625265Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:22:16.8658875Z Entering 'third_party/XNNPACK' 2025-12-04T09:22:16.8756022Z Entering 'third_party/aiter' 2025-12-04T09:22:16.8794186Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:22:16.8881238Z Entering 'third_party/benchmark' 2025-12-04T09:22:16.8911795Z Entering 'third_party/composable_kernel' 2025-12-04T09:22:16.9002777Z Entering 'third_party/cpp-httplib' 2025-12-04T09:22:16.9033177Z Entering 'third_party/cpuinfo' 2025-12-04T09:22:16.9069790Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:22:16.9102374Z Entering 'third_party/cutlass' 2025-12-04T09:22:16.9184780Z Entering 'third_party/fbgemm' 2025-12-04T09:22:16.9238172Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:22:16.9267760Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:22:16.9361323Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:22:16.9391807Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:22:16.9474255Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:22:16.9506040Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:22:16.9538497Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:22:16.9578657Z Entering 'third_party/flash-attention' 2025-12-04T09:22:16.9611100Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:22:16.9691054Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:22:16.9769175Z Entering 'third_party/flatbuffers' 2025-12-04T09:22:16.9822809Z Entering 'third_party/fmt' 2025-12-04T09:22:16.9854921Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:22:16.9889760Z Entering 'third_party/gloo' 2025-12-04T09:22:16.9914454Z Entering 'third_party/googletest' 2025-12-04T09:22:16.9950751Z Entering 'third_party/ideep' 2025-12-04T09:22:16.9982237Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:22:17.0051528Z Entering 'third_party/ittapi' 2025-12-04T09:22:17.0081079Z Entering 'third_party/kineto' 2025-12-04T09:22:17.0118545Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:22:17.0146687Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:22:17.0185440Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:22:17.0213187Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:22:17.0245130Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:22:17.0270392Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:22:17.0302394Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:22:17.0335879Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:22:17.0367969Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:22:17.0403545Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:22:17.0431359Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:22:17.0461386Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:17.0502807Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:17.0543170Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:22:17.0575939Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:22:17.0606117Z Entering 'third_party/kleidiai' 2025-12-04T09:22:17.0648077Z Entering 'third_party/mimalloc' 2025-12-04T09:22:17.0680199Z Entering 'third_party/nlohmann' 2025-12-04T09:22:17.0720018Z Entering 'third_party/onnx' 2025-12-04T09:22:17.0937584Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:22:17.0972867Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:22:17.1015872Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:22:17.1047436Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:22:17.1077338Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:22:17.1110010Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:22:17.1152801Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:22:17.1183693Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:22:17.1212356Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:22:17.1243635Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:17.1286544Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:17.1318448Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:22:17.1507392Z Entering 'third_party/pocketfft' 2025-12-04T09:22:17.1541266Z Entering 'third_party/protobuf' 2025-12-04T09:22:17.1602111Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:22:17.1630705Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:22:17.1667285Z Entering 'third_party/psimd' 2025-12-04T09:22:17.1698764Z Entering 'third_party/pthreadpool' 2025-12-04T09:22:17.1730550Z Entering 'third_party/pybind11' 2025-12-04T09:22:17.1765945Z Entering 'third_party/python-peachpy' 2025-12-04T09:22:17.1795436Z Entering 'third_party/sleef' 2025-12-04T09:22:17.1829478Z Entering 'third_party/tensorpipe' 2025-12-04T09:22:17.1863297Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:22:17.1899597Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:22:17.1924729Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:22:17.1959848Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:22:17.1990590Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:22:17.2115366Z Prepare all required actions 2025-12-04T09:22:17.2115773Z Getting action download info 2025-12-04T09:22:17.3746687Z ##[group]Run ./.github/actions/setup-linux 2025-12-04T09:22:17.3746897Z env: 2025-12-04T09:22:17.3747053Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:17.3747223Z ##[endgroup] 2025-12-04T09:22:17.3777959Z ##[group]Run set -euo pipefail 2025-12-04T09:22:17.3778196Z set -euo pipefail 2025-12-04T09:22:17.3778388Z function get_ec2_metadata() { 2025-12-04T09:22:17.3778631Z  # Pulled from instance metadata endpoint for EC2 2025-12-04T09:22:17.3779002Z  # see https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/instancedata-data-retrieval.html 2025-12-04T09:22:17.3779330Z  category=$1 2025-12-04T09:22:17.3779560Z  # If it is GCP runner (runner name contains gcp), do not run this 2025-12-04T09:22:17.3779815Z  runner_name_str=i-0c3ca63174fa64465 2025-12-04T09:22:17.3780093Z  if [[ -f /.inarc ]]; then 2025-12-04T09:22:17.3780305Z  echo "ARC Runner, no info on ec2 metadata" 2025-12-04T09:22:17.3780539Z  elif [[ $runner_name_str == *"gcp"* ]]; then 2025-12-04T09:22:17.3780811Z  echo "Runner is from Google Cloud Platform, No info on ec2 metadata" 2025-12-04T09:22:17.3781055Z  else 2025-12-04T09:22:17.3781531Z  curl -H "X-aws-ec2-metadata-token: $(curl -s -X PUT "http://169.254.169.254/latest/api/token" -H "X-aws-ec2-metadata-token-ttl-seconds: 30")" -fsSL "http://169.254.169.254/latest/meta-data/${category}" 2025-12-04T09:22:17.3782016Z  fi 2025-12-04T09:22:17.3782162Z } 2025-12-04T09:22:17.3782337Z echo "ami-id: $(get_ec2_metadata ami-id)" 2025-12-04T09:22:17.3782587Z echo "instance-id: $(get_ec2_metadata instance-id)" 2025-12-04T09:22:17.3782868Z echo "instance-type: $(get_ec2_metadata instance-type)" 2025-12-04T09:22:17.3783281Z echo "system info $(uname -a)" 2025-12-04T09:22:17.3787829Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:17.3788054Z env: 2025-12-04T09:22:17.3788209Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:17.3788370Z ##[endgroup] 2025-12-04T09:22:17.3969788Z ami-id: ami-08982f1c5bf93d976 2025-12-04T09:22:17.4066065Z instance-id: i-0c3ca63174fa64465 2025-12-04T09:22:17.4196897Z instance-type: m7i-flex.8xlarge 2025-12-04T09:22:17.4207863Z system info Linux ip-10-0-45-203.ec2.internal 6.1.150-174.273.amzn2023.x86_64 #1 SMP PREEMPT_DYNAMIC Tue Sep 9 12:21:26 UTC 2025 x86_64 x86_64 x86_64 GNU/Linux 2025-12-04T09:22:17.4223094Z ##[group]Run if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-12-04T09:22:17.4223394Z if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-12-04T09:22:17.4228035Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:17.4228262Z env: 2025-12-04T09:22:17.4228405Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:17.4228598Z ##[endgroup] 2025-12-04T09:22:17.4265991Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:22:17.4266507Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:22:17.4270620Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:17.4270859Z env: 2025-12-04T09:22:17.4271004Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:17.4271179Z ##[endgroup] 2025-12-04T09:22:17.4315690Z ##[group]Run if systemctl is-active --quiet docker; then 2025-12-04T09:22:17.4315971Z if systemctl is-active --quiet docker; then 2025-12-04T09:22:17.4316214Z  echo "Docker daemon is running..."; 2025-12-04T09:22:17.4316421Z else 2025-12-04T09:22:17.4316641Z  echo "Starting docker daemon..." && sudo systemctl start docker; 2025-12-04T09:22:17.4316908Z fi 2025-12-04T09:22:17.4320679Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:17.4320921Z env: 2025-12-04T09:22:17.4321071Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:17.4321252Z ##[endgroup] 2025-12-04T09:22:17.4409842Z Docker daemon is running... 2025-12-04T09:22:17.4437693Z ##[group]Run nick-fields/retry@v3.0.0 2025-12-04T09:22:17.4437895Z with: 2025-12-04T09:22:17.4438032Z shell: bash 2025-12-04T09:22:17.4438185Z timeout_minutes: 5 2025-12-04T09:22:17.4438348Z max_attempts: 3 2025-12-04T09:22:17.4438502Z retry_wait_seconds: 30 2025-12-04T09:22:17.4439778Z command: AWS_ACCOUNT_ID=$(aws sts get-caller-identity|grep Account|cut -f4 -d\") aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" # For LF Runners we need to make sure we also login to Meta's ECR docker registry too. META_AWS_ACCOUNT_ID=308535385114 if [ "$AWS_ACCOUNT_ID" != "$META_AWS_ACCOUNT_ID" ] ; then aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$META_AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" fi 2025-12-04T09:22:17.4440996Z polling_interval_seconds: 1 2025-12-04T09:22:17.4441179Z warning_on_retry: true 2025-12-04T09:22:17.4441351Z continue_on_error: false 2025-12-04T09:22:17.4441523Z env: 2025-12-04T09:22:17.4441688Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:17.4441992Z AWS_RETRY_MODE: standard 2025-12-04T09:22:17.4442182Z AWS_MAX_ATTEMPTS: 5 2025-12-04T09:22:17.4442382Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:22:17.4442589Z ##[endgroup] 2025-12-04T09:22:18.3117209Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:22:18.3117645Z Configure a credential helper to remove this warning. See 2025-12-04T09:22:18.3118011Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:22:18.3118539Z 2025-12-04T09:22:18.3118611Z Login Succeeded 2025-12-04T09:22:18.5119101Z Command completed after 1 attempt(s). 2025-12-04T09:22:18.5169324Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:22:18.5169649Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:22:18.5169921Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:22:18.5174724Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:18.5174945Z env: 2025-12-04T09:22:18.5175096Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:18.5175266Z ##[endgroup] 2025-12-04T09:22:18.5247699Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T09:22:18.5248033Z # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T09:22:18.5248289Z # shellcheck disable=SC2046 2025-12-04T09:22:18.5248505Z docker stop $(docker ps -q) || true 2025-12-04T09:22:18.5248720Z # Prune all of the docker images 2025-12-04T09:22:18.5248975Z docker system prune -af 2025-12-04T09:22:18.5252879Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:18.5253110Z env: 2025-12-04T09:22:18.5253263Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:18.5253423Z ##[endgroup] 2025-12-04T09:22:18.5669275Z "docker stop" requires at least 1 argument. 2025-12-04T09:22:18.5669586Z See 'docker stop --help'. 2025-12-04T09:22:18.5669705Z 2025-12-04T09:22:18.5669844Z Usage: docker stop [OPTIONS] CONTAINER [CONTAINER...] 2025-12-04T09:22:18.5670012Z 2025-12-04T09:22:18.5670101Z Stop one or more running containers 2025-12-04T09:22:18.5883658Z Total reclaimed space: 0B 2025-12-04T09:22:18.6004387Z ##[group]Run pytorch/test-infra/.github/actions/calculate-docker-image@main 2025-12-04T09:22:18.6004729Z with: 2025-12-04T09:22:18.6005327Z docker-image-name: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:18.6005911Z use-custom-docker-registry: true 2025-12-04T09:22:18.6006111Z docker-build-dir: .ci/docker 2025-12-04T09:22:18.6006306Z docker-build-script: ./build.sh 2025-12-04T09:22:18.6006496Z working-directory: . 2025-12-04T09:22:18.6006713Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:18.6006954Z force-push: false 2025-12-04T09:22:18.6007110Z env: 2025-12-04T09:22:18.6007253Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:18.6007426Z ##[endgroup] 2025-12-04T09:22:18.6020261Z ##[group]Run set -ex 2025-12-04T09:22:18.6020479Z set -ex 2025-12-04T09:22:18.6020652Z  2025-12-04T09:22:18.6020941Z # If the docker build directory or the build script doesn't exist, the action will 2025-12-04T09:22:18.6021324Z # gracefully return the docker image name as it is. Pulling docker image in Linux 2025-12-04T09:22:18.6021662Z # job could then download the pre-built image as usual 2025-12-04T09:22:18.6022070Z if [[ -d "${DOCKER_BUILD_DIR}" ]] && [[ -f "${DOCKER_BUILD_DIR}/${DOCKER_BUILD_SCRIPT}" ]] && [[ "${USE_CUSTOM_DOCKER_REGISTRY}" == "true" ]]; then 2025-12-04T09:22:18.6022434Z  echo "skip=false" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:18.6022641Z else 2025-12-04T09:22:18.6022816Z  echo "skip=true" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:18.6023088Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:18.6023330Z  2025-12-04T09:22:18.6023659Z  echo "Not using custom ECR registry. Either it was not requested or there is no Docker build script in the ${REPO_NAME} repo..." 2025-12-04T09:22:18.6024019Z  exit 0 2025-12-04T09:22:18.6024161Z fi 2025-12-04T09:22:18.6024305Z  2025-12-04T09:22:18.6024520Z if [[ "${DOCKER_IMAGE_NAME}" == *"${DOCKER_REGISTRY}/${REPO_NAME}"* ]]; then 2025-12-04T09:22:18.6024872Z  # The docker image name already includes the ECR prefix and tag, so we can just 2025-12-04T09:22:18.6025270Z  # use it as it is, but first let's extract the tag 2025-12-04T09:22:18.6025559Z  DOCKER_TAG=$(echo "${DOCKER_IMAGE_NAME}" | awk -F '[:,]' '{print $2}') 2025-12-04T09:22:18.6025860Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:18.6026141Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:18.6026379Z else 2025-12-04T09:22:18.6026560Z  if [[ "${DOCKER_IMAGE_NAME}" == *:* ]]; then 2025-12-04T09:22:18.6026797Z  CUSTOM_TAG_PREFIX=${DOCKER_IMAGE_NAME#*:} 2025-12-04T09:22:18.6027032Z  DOCKER_IMAGE_NAME=${DOCKER_IMAGE_NAME%%:*} 2025-12-04T09:22:18.6027237Z  fi 2025-12-04T09:22:18.6027510Z  DOCKER_TAG=${CUSTOM_TAG_PREFIX:+${CUSTOM_TAG_PREFIX}-}$(git rev-parse HEAD:"${DOCKER_BUILD_DIR}") 2025-12-04T09:22:18.6027853Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:18.6028218Z  echo "docker-image=${DOCKER_REGISTRY}/${REPO_NAME}/${DOCKER_IMAGE_NAME}:${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:18.6028611Z  echo "custom-tag-prefix=${CUSTOM_TAG_PREFIX}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:18.6028866Z fi 2025-12-04T09:22:18.6033488Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:18.6033722Z env: 2025-12-04T09:22:18.6033876Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:18.6034050Z REPO_NAME: pytorch 2025-12-04T09:22:18.6034670Z DOCKER_IMAGE_NAME: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:18.6035259Z DOCKER_BUILD_DIR: .ci/docker 2025-12-04T09:22:18.6035452Z DOCKER_BUILD_SCRIPT: ./build.sh 2025-12-04T09:22:18.6035691Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:18.6035944Z USE_CUSTOM_DOCKER_REGISTRY: true 2025-12-04T09:22:18.6036140Z CUSTOM_TAG_PREFIX: 2025-12-04T09:22:18.6036297Z ##[endgroup] 2025-12-04T09:22:18.6059091Z + [[ -d .ci/docker ]] 2025-12-04T09:22:18.6064394Z + [[ -f .ci/docker/./build.sh ]] 2025-12-04T09:22:18.6068866Z + [[ true == \t\r\u\e ]] 2025-12-04T09:22:18.6072959Z + echo skip=false 2025-12-04T09:22:18.6078677Z + [[ 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a == *\3\0\8\5\3\5\3\8\5\1\1\4\.\d\k\r\.\e\c\r\.\u\s\-\e\a\s\t\-\1\.\a\m\a\z\o\n\a\w\s\.\c\o\m\/\p\y\t\o\r\c\h* ]] 2025-12-04T09:22:18.6079732Z ++ echo 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:18.6080259Z ++ awk -F '[:,]' '{print $2}' 2025-12-04T09:22:18.6092200Z + DOCKER_TAG=pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:18.6100848Z + echo docker-tag=pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:18.6101721Z + echo docker-image=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:18.6117466Z ##[group]Run set +e 2025-12-04T09:22:18.6117674Z set +e 2025-12-04T09:22:18.6117833Z set -x 2025-12-04T09:22:18.6117987Z  2025-12-04T09:22:18.6118136Z login() { 2025-12-04T09:22:18.6118433Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-12-04T09:22:18.6118757Z } 2025-12-04T09:22:18.6118928Z  2025-12-04T09:22:18.6119076Z retry () { 2025-12-04T09:22:18.6119265Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-12-04T09:22:18.6119464Z } 2025-12-04T09:22:18.6119608Z  2025-12-04T09:22:18.6119862Z retry login "${DOCKER_REGISTRY}" 2025-12-04T09:22:18.6120055Z  2025-12-04T09:22:18.6120209Z START_TIME=$(date +%s) 2025-12-04T09:22:18.6120410Z # Wait up to 120 minutes 2025-12-04T09:22:18.6120647Z while [[ $(( $(date +%s) - 7200 )) -lt $START_TIME ]]; do 2025-12-04T09:22:18.6120949Z  # Check if image already exists, if it does then skip building it 2025-12-04T09:22:18.6121251Z  if docker manifest inspect "${DOCKER_IMAGE}"; then 2025-12-04T09:22:18.6121482Z  exit 0 2025-12-04T09:22:18.6121636Z  fi 2025-12-04T09:22:18.6122022Z  2025-12-04T09:22:18.6122313Z  # NB: This flag is used by Docker build workflow to push the image to ECR, so we can 2025-12-04T09:22:18.6122749Z  # use this to differentiate between the Docker build and regular build jobs. For the 2025-12-04T09:22:18.6123192Z  # latter, it will wait for the Docker images to become available before continuing 2025-12-04T09:22:18.6123540Z  if [ "${DOCKER_PUSH:-false}" == "true" ]; then 2025-12-04T09:22:18.6123794Z  # It's a Docker build job, let's build the image 2025-12-04T09:22:18.6124003Z  break 2025-12-04T09:22:18.6124160Z  else 2025-12-04T09:22:18.6124376Z  # It's a regular build job, wait for the image to become available 2025-12-04T09:22:18.6124616Z  sleep 300 2025-12-04T09:22:18.6124780Z  fi 2025-12-04T09:22:18.6124927Z done 2025-12-04T09:22:18.6125065Z  2025-12-04T09:22:18.6125288Z # NB: This part requires a full checkout. Otherwise, the merge base will 2025-12-04T09:22:18.6125680Z # be empty. The default action would be to continue rebuild the image 2025-12-04T09:22:18.6125992Z if [[ "$BASE_REVISION" = "$(git rev-parse HEAD)" ]]; then 2025-12-04T09:22:18.6126269Z  # if we're on the base branch then use the parent commit 2025-12-04T09:22:18.6126526Z  MERGE_BASE=$(git rev-parse HEAD~) 2025-12-04T09:22:18.6126728Z else 2025-12-04T09:22:18.6126935Z  # otherwise we're on a PR, so use the most recent base commit 2025-12-04T09:22:18.6127225Z  MERGE_BASE=$(git merge-base HEAD "$BASE_REVISION") 2025-12-04T09:22:18.6127448Z fi 2025-12-04T09:22:18.6127593Z  2025-12-04T09:22:18.6127747Z if [[ -z "${MERGE_BASE}" ]]; then 2025-12-04T09:22:18.6127976Z  echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:18.6128186Z  2025-12-04T09:22:18.6128463Z  echo "Finding merge base only works with full checkout, please set fetch-depth to 0, continuing ..." 2025-12-04T09:22:18.6128785Z  exit 0 2025-12-04T09:22:18.6128940Z fi 2025-12-04T09:22:18.6129079Z  2025-12-04T09:22:18.6129280Z if ! git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}"; then 2025-12-04T09:22:18.6129674Z  echo "Directory '${DOCKER_BUILD_DIR}' not found in commit $MERGE_BASE, you should rebase onto a more recent commit" 2025-12-04T09:22:18.6130014Z  exit 1 2025-12-04T09:22:18.6130858Z fi 2025-12-04T09:22:18.6131206Z  2025-12-04T09:22:18.6131483Z PREVIOUS_DOCKER_TAG=$(git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}") 2025-12-04T09:22:18.6131892Z # If no image exists but the hash is the same as the previous hash then we should error out here 2025-12-04T09:22:18.6132275Z if [[ "${PREVIOUS_DOCKER_TAG}" == "${DOCKER_TAG}" ]]; then 2025-12-04T09:22:18.6132787Z  echo "WARNING: Something has gone wrong and the previous image isn't available for the merge-base of your branch" 2025-12-04T09:22:18.6133283Z  echo " Will re-build docker image to store in local cache, TTS may be longer" 2025-12-04T09:22:18.6133601Z fi 2025-12-04T09:22:18.6133747Z  2025-12-04T09:22:18.6133942Z echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:18.6138395Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:18.6138618Z env: 2025-12-04T09:22:18.6138769Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:18.6138944Z DOCKER_BUILD_DIR: .ci/docker 2025-12-04T09:22:18.6139162Z BASE_REVISION: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:22:18.6139719Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:18.6140387Z DOCKER_TAG: pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:18.6140812Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:18.6141036Z DOCKER_PUSH: 2025-12-04T09:22:18.6141189Z ##[endgroup] 2025-12-04T09:22:18.6160801Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:18.6161138Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:18.6166323Z + aws ecr get-login-password --region us-east-1 2025-12-04T09:22:18.6166892Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:19.0269801Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:22:19.0275761Z Configure a credential helper to remove this warning. See 2025-12-04T09:22:19.0276084Z Login Succeeded 2025-12-04T09:22:19.0277336Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:22:19.0277601Z 2025-12-04T09:22:19.0292012Z ++ date +%s 2025-12-04T09:22:19.0301038Z + START_TIME=1764840139 2025-12-04T09:22:19.0306073Z ++ date +%s 2025-12-04T09:22:19.0308778Z + [[ 1764832939 -lt 1764840139 ]] 2025-12-04T09:22:19.0309449Z + docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:19.2490005Z { 2025-12-04T09:22:19.2493874Z "schemaVersion": 2, 2025-12-04T09:22:19.2498400Z "mediaType": "application/vnd.docker.distribution.manifest.v2+json", 2025-12-04T09:22:19.2498811Z "config": { 2025-12-04T09:22:19.2499131Z "mediaType": "application/vnd.docker.container.image.v1+json", 2025-12-04T09:22:19.2499428Z "size": 33581, 2025-12-04T09:22:19.2499826Z "digest": "sha256:40bd34edc3497e05e2150804a15313235b6426bcc4b8c19c03c3daf5524be913" 2025-12-04T09:22:19.2500249Z }, 2025-12-04T09:22:19.2500461Z "layers": [ 2025-12-04T09:22:19.2500643Z { 2025-12-04T09:22:19.2500876Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2501126Z "size": 30447951, 2025-12-04T09:22:19.2501408Z "digest": "sha256:63e5bc7682b85ae57a1221210f64d62e7a90b0a30f19af4ca734b8242ae49d63" 2025-12-04T09:22:19.2501672Z }, 2025-12-04T09:22:19.2501799Z { 2025-12-04T09:22:19.2502003Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2502242Z "size": 1554, 2025-12-04T09:22:19.2502496Z "digest": "sha256:6dc15eca51381c13be16385052dd2378ab1dce5fb77f7e7bceab34ed72e6e0e5" 2025-12-04T09:22:19.2502763Z }, 2025-12-04T09:22:19.2502898Z { 2025-12-04T09:22:19.2503167Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2503508Z "size": 313274856, 2025-12-04T09:22:19.2503863Z "digest": "sha256:459f4df18f07c473cb662e8033f9dab9f68c9ac71608767ad055999fae6b31e2" 2025-12-04T09:22:19.2504226Z }, 2025-12-04T09:22:19.2504407Z { 2025-12-04T09:22:19.2504702Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2505064Z "size": 787, 2025-12-04T09:22:19.2505427Z "digest": "sha256:82108541691974f1cc469d0ec89a048191b8de4dc36bf9a0184a4bd50120b69d" 2025-12-04T09:22:19.2505803Z }, 2025-12-04T09:22:19.2505962Z { 2025-12-04T09:22:19.2506208Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2506453Z "size": 106, 2025-12-04T09:22:19.2506692Z "digest": "sha256:3738646b3d92a3f1b6529ba588c00216c306c997ad77730f207c7107c2e5aff3" 2025-12-04T09:22:19.2507240Z }, 2025-12-04T09:22:19.2507413Z { 2025-12-04T09:22:19.2507622Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2507852Z "size": 704, 2025-12-04T09:22:19.2508098Z "digest": "sha256:aa9a10a37b3945f69568695fa4a9321333654a69dd44c55eff4e26bf9bffcad1" 2025-12-04T09:22:19.2508410Z }, 2025-12-04T09:22:19.2508535Z { 2025-12-04T09:22:19.2508799Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2509042Z "size": 1218, 2025-12-04T09:22:19.2509293Z "digest": "sha256:55bac08f3e18897f61d3e7f28ada6547b05b3c36f559a882656834d43de61478" 2025-12-04T09:22:19.2509561Z }, 2025-12-04T09:22:19.2509690Z { 2025-12-04T09:22:19.2509890Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2510131Z "size": 485, 2025-12-04T09:22:19.2510377Z "digest": "sha256:acf6468f6aadc3613e3b8fe73c016ce8d6f47f01e0cfee3a1015d0335c24d691" 2025-12-04T09:22:19.2510653Z }, 2025-12-04T09:22:19.2510777Z { 2025-12-04T09:22:19.2510965Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2511205Z "size": 110362453, 2025-12-04T09:22:19.2511456Z "digest": "sha256:c0c31a0e69b798875a8d2c05d01a6861b2fe623e1970bf7faf822e7df6691d2c" 2025-12-04T09:22:19.2511711Z }, 2025-12-04T09:22:19.2511836Z { 2025-12-04T09:22:19.2512032Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2512261Z "size": 4961, 2025-12-04T09:22:19.2512505Z "digest": "sha256:0f7d0dc70d2d944cc1c8a592f051d3f28d2b062c79166713fb7cc9225b4c97c0" 2025-12-04T09:22:19.2512770Z }, 2025-12-04T09:22:19.2512981Z { 2025-12-04T09:22:19.2513183Z + exit 0 2025-12-04T09:22:19.2513390Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2513631Z "size": 1756, 2025-12-04T09:22:19.2513919Z "digest": "sha256:58c4d15d3bc18a789eb0d1a13f68df0b1d6cea66697fe63adf11b53fec74ba18" 2025-12-04T09:22:19.2514199Z }, 2025-12-04T09:22:19.2514318Z { 2025-12-04T09:22:19.2514522Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2514764Z "size": 724, 2025-12-04T09:22:19.2515009Z "digest": "sha256:2f1cc47b61e1ebefa16f8ee12c5a1df77b109a84a8e8f425e7d65ec51cafdb93" 2025-12-04T09:22:19.2515282Z }, 2025-12-04T09:22:19.2515408Z { 2025-12-04T09:22:19.2515604Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2515836Z "size": 543, 2025-12-04T09:22:19.2516082Z "digest": "sha256:d2472a74103d877e6532d6409a2ac9d54db6672b5d3002be011fdbbd8d360c49" 2025-12-04T09:22:19.2516347Z }, 2025-12-04T09:22:19.2516466Z { 2025-12-04T09:22:19.2516664Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2516905Z "size": 3353192867, 2025-12-04T09:22:19.2517153Z "digest": "sha256:d1fba72936888ea8f8e4c7ad015bfcf3285017b95f3d72795d5b6a0b0ded36fc" 2025-12-04T09:22:19.2517423Z }, 2025-12-04T09:22:19.2517553Z { 2025-12-04T09:22:19.2517741Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2517981Z "size": 32, 2025-12-04T09:22:19.2518232Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:19.2518500Z }, 2025-12-04T09:22:19.2518617Z { 2025-12-04T09:22:19.2518814Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2519085Z "size": 397, 2025-12-04T09:22:19.2535842Z "digest": "sha256:6a9b03ce41a77be16366fba2c6ca0aabefe42ee4879ab5d088b312970e830770" 2025-12-04T09:22:19.2536148Z }, 2025-12-04T09:22:19.2536290Z { 2025-12-04T09:22:19.2536525Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2536783Z "size": 237420, 2025-12-04T09:22:19.2537046Z "digest": "sha256:3519ac15be7978a14201f5d431b6472ed30382cde8628c6401c9db98ccd1a871" 2025-12-04T09:22:19.2537321Z }, 2025-12-04T09:22:19.2537451Z { 2025-12-04T09:22:19.2537840Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2538097Z "size": 231, 2025-12-04T09:22:19.2538359Z "digest": "sha256:dabb51b819a06823e5a1c8d4f9e29b689f38b19ee1e90da40075bd1593b8d0fb" 2025-12-04T09:22:19.2538634Z }, 2025-12-04T09:22:19.2538775Z { 2025-12-04T09:22:19.2538980Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2539215Z "size": 2967944, 2025-12-04T09:22:19.2539473Z "digest": "sha256:8f36833a24d02d47bf18bdc573adbb45afb8f5f06886da9bd671a1a33e3007bd" 2025-12-04T09:22:19.2539745Z }, 2025-12-04T09:22:19.2539863Z { 2025-12-04T09:22:19.2540121Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2540363Z "size": 1472, 2025-12-04T09:22:19.2540611Z "digest": "sha256:ab53c5a853a52de13f11fbb026aaf469754adbe52b02e96f4fbdc05916585244" 2025-12-04T09:22:19.2540883Z }, 2025-12-04T09:22:19.2541009Z { 2025-12-04T09:22:19.2541206Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2541445Z "size": 482, 2025-12-04T09:22:19.2541692Z "digest": "sha256:025a0e5e6ac19cb2d27c9423a5d019a34ea7d4004d426d6b890693ccfe9f54d3" 2025-12-04T09:22:19.2541959Z }, 2025-12-04T09:22:19.2542079Z { 2025-12-04T09:22:19.2542283Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2542525Z "size": 200, 2025-12-04T09:22:19.2542764Z "digest": "sha256:ce3394c8f2109f4538f6e7f2c39f685d0ba028aacac33ddacbb8ef9fd349702e" 2025-12-04T09:22:19.2543034Z }, 2025-12-04T09:22:19.2543160Z { 2025-12-04T09:22:19.2543349Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2543663Z "size": 608, 2025-12-04T09:22:19.2543920Z "digest": "sha256:a5c3888c3a0c5812efbfe307d18de826363900fa8d8097c1ee84a72630aa067b" 2025-12-04T09:22:19.2544196Z }, 2025-12-04T09:22:19.2544318Z { 2025-12-04T09:22:19.2544519Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2544764Z "size": 225, 2025-12-04T09:22:19.2545009Z "digest": "sha256:ed902a3a4e3b916aa869b612baf9cd36745b671b6781d19e3b4c3fadc2f513f5" 2025-12-04T09:22:19.2545290Z }, 2025-12-04T09:22:19.2545428Z { 2025-12-04T09:22:19.2545619Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2545859Z "size": 829, 2025-12-04T09:22:19.2546105Z "digest": "sha256:b4e1efca22beb475459a14f4a571a2512c2b1ed4b24e66bd3f82f1681dc5b1bc" 2025-12-04T09:22:19.2546370Z }, 2025-12-04T09:22:19.2546495Z { 2025-12-04T09:22:19.2546690Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2546922Z "size": 32, 2025-12-04T09:22:19.2547173Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:19.2547444Z }, 2025-12-04T09:22:19.2547567Z { 2025-12-04T09:22:19.2547756Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2547994Z "size": 104, 2025-12-04T09:22:19.2548240Z "digest": "sha256:79fcfd297d9fcfe2810595fc2c6ba503293b59046bc192bb0620a64d9bdff778" 2025-12-04T09:22:19.2548503Z }, 2025-12-04T09:22:19.2548630Z { 2025-12-04T09:22:19.2548827Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2549059Z "size": 1495, 2025-12-04T09:22:19.2549305Z "digest": "sha256:7bd89c134b49f83cfb0725df3615556b60af380b932186c5abd001232b4597cf" 2025-12-04T09:22:19.2549572Z }, 2025-12-04T09:22:19.2549688Z { 2025-12-04T09:22:19.2549884Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2550130Z "size": 458790783, 2025-12-04T09:22:19.2550387Z "digest": "sha256:b7d40b4fd1b9375cfd45d5c593ec0cccc21c80d667ea79f4effb09f04b0705b0" 2025-12-04T09:22:19.2550666Z }, 2025-12-04T09:22:19.2550790Z { 2025-12-04T09:22:19.2550987Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2551219Z "size": 163, 2025-12-04T09:22:19.2551466Z "digest": "sha256:dd1cdd87320d29d4ca09686ab00b76a396efcc6f3ea6d0bfd1f7922e46336ca6" 2025-12-04T09:22:19.2551779Z }, 2025-12-04T09:22:19.2551898Z { 2025-12-04T09:22:19.2552099Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2552339Z "size": 347, 2025-12-04T09:22:19.2552579Z "digest": "sha256:c21a6e1cd03a718ebfb927ab114ebbbee0b57a93908d0265d24b02cf34c0fee3" 2025-12-04T09:22:19.2552852Z }, 2025-12-04T09:22:19.2552982Z { 2025-12-04T09:22:19.2553175Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2553415Z "size": 32, 2025-12-04T09:22:19.2553664Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:19.2553933Z }, 2025-12-04T09:22:19.2554053Z { 2025-12-04T09:22:19.2554252Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2554493Z "size": 106, 2025-12-04T09:22:19.2554739Z "digest": "sha256:0aa3ea0b275426eac8eff6b26ba17c168ae70d1baa5a7845e61d41bd01a9ff43" 2025-12-04T09:22:19.2555019Z }, 2025-12-04T09:22:19.2555148Z { 2025-12-04T09:22:19.2555341Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2555584Z "size": 426, 2025-12-04T09:22:19.2555833Z "digest": "sha256:74e0bdbb05d31425d7a2f40ff0cb2423abaf90f1ed56c4b69f43e8d92c569379" 2025-12-04T09:22:19.2556104Z }, 2025-12-04T09:22:19.2556234Z { 2025-12-04T09:22:19.2556437Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2556678Z "size": 19309400, 2025-12-04T09:22:19.2556947Z "digest": "sha256:ba3aee7dcf03efbd1e7b1611c8f65f385c94b2572d15c099a1535f27b72509f7" 2025-12-04T09:22:19.2557288Z }, 2025-12-04T09:22:19.2557426Z { 2025-12-04T09:22:19.2557618Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2557864Z "size": 108, 2025-12-04T09:22:19.2558118Z "digest": "sha256:8c8d59b8759d40399b07b019a46fc05b9ccabe2b9276f49eea203fd6908ff334" 2025-12-04T09:22:19.2558386Z }, 2025-12-04T09:22:19.2558516Z { 2025-12-04T09:22:19.2558743Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2558979Z "size": 827, 2025-12-04T09:22:19.2559222Z "digest": "sha256:3c960865867b66327e316bf682adabd1867dd9a41cc0aee1507ee8e35fa614a9" 2025-12-04T09:22:19.2559486Z }, 2025-12-04T09:22:19.2559608Z { 2025-12-04T09:22:19.2559806Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2560050Z "size": 724, 2025-12-04T09:22:19.2560295Z "digest": "sha256:2f1cc47b61e1ebefa16f8ee12c5a1df77b109a84a8e8f425e7d65ec51cafdb93" 2025-12-04T09:22:19.2560573Z }, 2025-12-04T09:22:19.2560700Z { 2025-12-04T09:22:19.2560901Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2561137Z "size": 149, 2025-12-04T09:22:19.2561382Z "digest": "sha256:074db79e3832af094ddad6e5c5649347b62e8ba0365229feced3ed0263a0c611" 2025-12-04T09:22:19.2561656Z }, 2025-12-04T09:22:19.2561869Z { 2025-12-04T09:22:19.2562079Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2562326Z "size": 137, 2025-12-04T09:22:19.2562589Z "digest": "sha256:226748f4ff23ba2aef33619c3a639fb44865d5f4515c4b45d4196f4e0e77b187" 2025-12-04T09:22:19.2562902Z }, 2025-12-04T09:22:19.2563046Z { 2025-12-04T09:22:19.2563252Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2563501Z "size": 140, 2025-12-04T09:22:19.2563746Z "digest": "sha256:75d817336f1df7579f5b4be29625c53d594cb78c776f5454cc69438bb1bb0dc9" 2025-12-04T09:22:19.2564017Z }, 2025-12-04T09:22:19.2564139Z { 2025-12-04T09:22:19.2564342Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2564590Z "size": 18890247472, 2025-12-04T09:22:19.2564842Z "digest": "sha256:850870a17b94a4d7c5e55b287cc0b1ddd0982845e11f055765dc358277cdfc54" 2025-12-04T09:22:19.2565113Z }, 2025-12-04T09:22:19.2565239Z { 2025-12-04T09:22:19.2565486Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2565728Z "size": 222, 2025-12-04T09:22:19.2565973Z "digest": "sha256:e6225129924f0a66acaf35efd792901351da58116629962669d9c3acaadbab6e" 2025-12-04T09:22:19.2566236Z }, 2025-12-04T09:22:19.2566362Z { 2025-12-04T09:22:19.2566563Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2566798Z "size": 255, 2025-12-04T09:22:19.2567111Z "digest": "sha256:5cd6b9ec5c3c4a1cf2e2e7a1794aaae173eb5a291b9899bb733e5f36d866fab0" 2025-12-04T09:22:19.2567392Z }, 2025-12-04T09:22:19.2567522Z { 2025-12-04T09:22:19.2567717Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2567959Z "size": 32, 2025-12-04T09:22:19.2568212Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:19.2568481Z }, 2025-12-04T09:22:19.2568611Z { 2025-12-04T09:22:19.2568809Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2569047Z "size": 106, 2025-12-04T09:22:19.2569297Z "digest": "sha256:613d2b1f3db875d1d695081b44aefbf94eeb7815d794b28dd17ec623d3f9dad5" 2025-12-04T09:22:19.2569574Z }, 2025-12-04T09:22:19.2569693Z { 2025-12-04T09:22:19.2569889Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2570130Z "size": 312293443, 2025-12-04T09:22:19.2570379Z "digest": "sha256:388169fffe8a06396f4e8d6fc51d07a83e08f04836995ae28c2e13960cb89fab" 2025-12-04T09:22:19.2570647Z }, 2025-12-04T09:22:19.2570773Z { 2025-12-04T09:22:19.2570971Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2571287Z "size": 6628443345, 2025-12-04T09:22:19.2571544Z "digest": "sha256:6ca498e78b7b821a815bc7c14f057d52f7d9ce8dda8bca0945e6f522c7d208e1" 2025-12-04T09:22:19.2571815Z }, 2025-12-04T09:22:19.2571946Z { 2025-12-04T09:22:19.2572143Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2572376Z "size": 129, 2025-12-04T09:22:19.2572623Z "digest": "sha256:952b753ec7be39e1967f95afb87a880712b6425610adcda2afe3b2211cc8de12" 2025-12-04T09:22:19.2572892Z }, 2025-12-04T09:22:19.2573022Z { 2025-12-04T09:22:19.2573214Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2573454Z "size": 880, 2025-12-04T09:22:19.2573701Z "digest": "sha256:5b6f6de4bdd1c96aa8704c326fb27fa7465ebf388f0748ee56d9f521bb6b6697" 2025-12-04T09:22:19.2573963Z }, 2025-12-04T09:22:19.2574087Z { 2025-12-04T09:22:19.2574282Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2574512Z "size": 724, 2025-12-04T09:22:19.2574765Z "digest": "sha256:2f1cc47b61e1ebefa16f8ee12c5a1df77b109a84a8e8f425e7d65ec51cafdb93" 2025-12-04T09:22:19.2575035Z }, 2025-12-04T09:22:19.2575150Z { 2025-12-04T09:22:19.2575347Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2575583Z "size": 139, 2025-12-04T09:22:19.2575834Z "digest": "sha256:fa23d9952f3cea91af7b5ffce93e64d90d98aefb838aa8a1fe4a6b40cd0eab91" 2025-12-04T09:22:19.2576095Z }, 2025-12-04T09:22:19.2576217Z { 2025-12-04T09:22:19.2576412Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2576642Z "size": 32, 2025-12-04T09:22:19.2576887Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:19.2577151Z }, 2025-12-04T09:22:19.2577267Z { 2025-12-04T09:22:19.2577460Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2577697Z "size": 161, 2025-12-04T09:22:19.2577946Z "digest": "sha256:c2aa9ea4d09ac3edb41e48cdc892fafe72a5cddc98ebdf67d978c6f8d63cd7d2" 2025-12-04T09:22:19.2578221Z }, 2025-12-04T09:22:19.2578345Z { 2025-12-04T09:22:19.2578532Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2578768Z "size": 1011, 2025-12-04T09:22:19.2579018Z "digest": "sha256:a9bbc9e426d367e15c7f0c1faae6417526e219fa8188ae17b9e6c2c3b8083bdc" 2025-12-04T09:22:19.2579332Z }, 2025-12-04T09:22:19.2579450Z { 2025-12-04T09:22:19.2579644Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2579881Z "size": 724, 2025-12-04T09:22:19.2580122Z "digest": "sha256:2f1cc47b61e1ebefa16f8ee12c5a1df77b109a84a8e8f425e7d65ec51cafdb93" 2025-12-04T09:22:19.2580393Z }, 2025-12-04T09:22:19.2580518Z { 2025-12-04T09:22:19.2580705Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2580943Z "size": 135, 2025-12-04T09:22:19.2581188Z "digest": "sha256:82f2371f6dc2c2797baa60e672fd78bb04909cd0248c52cae058a7fd8f215a4f" 2025-12-04T09:22:19.2581451Z }, 2025-12-04T09:22:19.2581580Z { 2025-12-04T09:22:19.2581775Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2582000Z "size": 32, 2025-12-04T09:22:19.2582243Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:19.2582511Z }, 2025-12-04T09:22:19.2582635Z { 2025-12-04T09:22:19.2582824Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2583058Z "size": 158, 2025-12-04T09:22:19.2583302Z "digest": "sha256:c87301921afac5ba5745add04b44f57bf31c551bf48a2574d4424dab64c25ce2" 2025-12-04T09:22:19.2583559Z }, 2025-12-04T09:22:19.2583684Z { 2025-12-04T09:22:19.2583877Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2584107Z "size": 603, 2025-12-04T09:22:19.2584354Z "digest": "sha256:384d566a822f4defa0ecf9c7d37b6d7a5fc5eaab415fe1d3cc02b190ddc19e71" 2025-12-04T09:22:19.2584626Z }, 2025-12-04T09:22:19.2584783Z { 2025-12-04T09:22:19.2584983Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2585222Z "size": 724, 2025-12-04T09:22:19.2585470Z "digest": "sha256:2f1cc47b61e1ebefa16f8ee12c5a1df77b109a84a8e8f425e7d65ec51cafdb93" 2025-12-04T09:22:19.2585735Z }, 2025-12-04T09:22:19.2585861Z { 2025-12-04T09:22:19.2586058Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2586291Z "size": 155, 2025-12-04T09:22:19.2586536Z "digest": "sha256:32ddcdd5dd6920acb12ca95671a164346308ec7743afbe8a333142733c42bafd" 2025-12-04T09:22:19.2586807Z }, 2025-12-04T09:22:19.2586926Z { 2025-12-04T09:22:19.2587125Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2587362Z "size": 32, 2025-12-04T09:22:19.2587602Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:19.2587870Z }, 2025-12-04T09:22:19.2587994Z { 2025-12-04T09:22:19.2588185Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2588424Z "size": 188, 2025-12-04T09:22:19.2588670Z "digest": "sha256:a116a32ceaa15bbd55b0f4281fb0040de221c62e29659b23dd71ecd2382a969a" 2025-12-04T09:22:19.2588935Z }, 2025-12-04T09:22:19.2589054Z { 2025-12-04T09:22:19.2589245Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2589485Z "size": 1371, 2025-12-04T09:22:19.2589727Z "digest": "sha256:f6fbbbe6067e383cadb530c3f53b113e9826843fb2348ea1f9a000a4790fa822" 2025-12-04T09:22:19.2589998Z }, 2025-12-04T09:22:19.2590122Z { 2025-12-04T09:22:19.2590309Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2590549Z "size": 32, 2025-12-04T09:22:19.2590792Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:19.2591053Z }, 2025-12-04T09:22:19.2591179Z { 2025-12-04T09:22:19.2591379Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2591610Z "size": 137, 2025-12-04T09:22:19.2591847Z "digest": "sha256:648018658875ac3356b277abe3d98df9a8b7a0345f131598c07f5e9f9318b740" 2025-12-04T09:22:19.2592110Z }, 2025-12-04T09:22:19.2592233Z { 2025-12-04T09:22:19.2592422Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2592703Z "size": 528, 2025-12-04T09:22:19.2592951Z "digest": "sha256:5e2f7eac20ad5128eebe0ba7dbd08111d28e7f65c26c7fea7cd6dc7a2c0725b9" 2025-12-04T09:22:19.2593218Z }, 2025-12-04T09:22:19.2593344Z { 2025-12-04T09:22:19.2593539Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2593767Z "size": 32, 2025-12-04T09:22:19.2594011Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:19.2594274Z }, 2025-12-04T09:22:19.2594391Z { 2025-12-04T09:22:19.2594586Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2594828Z "size": 104, 2025-12-04T09:22:19.2595068Z "digest": "sha256:0d1243af7593a042ff443492a4cac70f817dc8c56ac45a0c138d102f40c1cc07" 2025-12-04T09:22:19.2595326Z }, 2025-12-04T09:22:19.2595449Z { 2025-12-04T09:22:19.2595643Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2595873Z "size": 435, 2025-12-04T09:22:19.2596117Z "digest": "sha256:1f6d5b941ea1f6f07c190f5ca1d4a5172e4d696ad597ae58dcd37b297f60b556" 2025-12-04T09:22:19.2596380Z }, 2025-12-04T09:22:19.2596495Z { 2025-12-04T09:22:19.2596691Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2596926Z "size": 32, 2025-12-04T09:22:19.2597162Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:19.2597417Z }, 2025-12-04T09:22:19.2597534Z { 2025-12-04T09:22:19.2597721Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2597955Z "size": 107, 2025-12-04T09:22:19.2598237Z "digest": "sha256:4444f70d73cefec9f161242e9def76078af5abc7a3ad3e5d6e7d4f1efed07939" 2025-12-04T09:22:19.2598511Z }, 2025-12-04T09:22:19.2598629Z { 2025-12-04T09:22:19.2598820Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2599051Z "size": 1896, 2025-12-04T09:22:19.2599288Z "digest": "sha256:bf45550ddbad2524852930466f963e432429f2b86f7f6ccf0415fc56980004ce" 2025-12-04T09:22:19.2599549Z }, 2025-12-04T09:22:19.2599671Z { 2025-12-04T09:22:19.2599858Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2600097Z "size": 245588924, 2025-12-04T09:22:19.2600349Z "digest": "sha256:30fff4f4bad325a3a2d91463ca6702c411392e5b0646a5cd4f47fcae56c55639" 2025-12-04T09:22:19.2600614Z }, 2025-12-04T09:22:19.2600740Z { 2025-12-04T09:22:19.2600936Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2601163Z "size": 106, 2025-12-04T09:22:19.2601409Z "digest": "sha256:cca4db1ba155a3afd1fd645cd3223e9a4de9c785f8559387f21da187672d9e9e" 2025-12-04T09:22:19.2601700Z }, 2025-12-04T09:22:19.2602045Z { 2025-12-04T09:22:19.2602268Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2602552Z "size": 165, 2025-12-04T09:22:19.2602840Z "digest": "sha256:f0ffa379f4eba2f3ddc323f45f226f39aed5a7e93880884a97bd4d6a91864dc1" 2025-12-04T09:22:19.2603148Z }, 2025-12-04T09:22:19.2603274Z { 2025-12-04T09:22:19.2603470Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2603728Z "size": 7942, 2025-12-04T09:22:19.2604002Z "digest": "sha256:df645e678f95db120f96e4c4c66400d315ed9214d9085216b417348cf682d43a" 2025-12-04T09:22:19.2604307Z }, 2025-12-04T09:22:19.2604438Z { 2025-12-04T09:22:19.2604659Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2604941Z "size": 8076, 2025-12-04T09:22:19.2605221Z "digest": "sha256:52ad6c17d10308710fb2dc855f911246b7a369ccf5c3142cb3ad7fb182f708b8" 2025-12-04T09:22:19.2605514Z }, 2025-12-04T09:22:19.2605653Z { 2025-12-04T09:22:19.2605873Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2606141Z "size": 302, 2025-12-04T09:22:19.2606407Z "digest": "sha256:34408374e32c70dbcfd7e0516edaa8698c25e6e3dbcd920978f2b3a801bbe659" 2025-12-04T09:22:19.2606748Z }, 2025-12-04T09:22:19.2606876Z { 2025-12-04T09:22:19.2607090Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2607361Z "size": 32, 2025-12-04T09:22:19.2607627Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:19.2607932Z }, 2025-12-04T09:22:19.2608068Z { 2025-12-04T09:22:19.2608281Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2608562Z "size": 108, 2025-12-04T09:22:19.2608802Z "digest": "sha256:27803b661d9e2cb1ce49c0a100ce824fbcbdd1f3a3ceda4ed5affbcb07ac3be4" 2025-12-04T09:22:19.2609069Z }, 2025-12-04T09:22:19.2609186Z { 2025-12-04T09:22:19.2609378Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2609614Z "size": 54145699, 2025-12-04T09:22:19.2609858Z "digest": "sha256:07dafc893dea952d7677124b42e90b29b167fabe81e8a9d8b3fb4b0aa11b66ec" 2025-12-04T09:22:19.2610123Z }, 2025-12-04T09:22:19.2610245Z { 2025-12-04T09:22:19.2610436Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:19.2610669Z "size": 32, 2025-12-04T09:22:19.2610910Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:19.2611169Z } 2025-12-04T09:22:19.2611292Z ] 2025-12-04T09:22:19.2611421Z } 2025-12-04T09:22:19.2633343Z ##[group]Run set -eux 2025-12-04T09:22:19.2633545Z set -eux 2025-12-04T09:22:19.2633801Z # It's ok if this steps fails, it would then be an anonymous user like what we used to have 2025-12-04T09:22:19.2634531Z aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token | jq --raw-output '.SecretString' | jq -r .docker_hub_readonly_token | docker login --username pytorchbot --password-stdin || true 2025-12-04T09:22:19.2639274Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:19.2639516Z env: 2025-12-04T09:22:19.2639681Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:19.2639864Z ##[endgroup] 2025-12-04T09:22:19.2667235Z + aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token 2025-12-04T09:22:19.2669999Z + docker login --username pytorchbot --password-stdin 2025-12-04T09:22:19.2670373Z + jq -r .docker_hub_readonly_token 2025-12-04T09:22:19.2676163Z + jq --raw-output .SecretString 2025-12-04T09:22:19.7077428Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:22:19.7077774Z Login Succeeded 2025-12-04T09:22:19.7078009Z Configure a credential helper to remove this warning. See 2025-12-04T09:22:19.7078405Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:22:19.7078635Z 2025-12-04T09:22:19.7152705Z ##[group]Run tag=${ECR_DOCKER_IMAGE##*:} 2025-12-04T09:22:19.7152955Z tag=${ECR_DOCKER_IMAGE##*:} 2025-12-04T09:22:19.7153207Z echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}" 2025-12-04T09:22:19.7157499Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:19.7157730Z env: 2025-12-04T09:22:19.7157886Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:19.7158391Z ECR_DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:19.7158886Z ##[endgroup] 2025-12-04T09:22:19.7181342Z docker pull ghcr.io/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:19.7214864Z ##[group]Run pytorch/test-infra/.github/actions/pull-docker-image@main 2025-12-04T09:22:19.7215157Z with: 2025-12-04T09:22:19.7215731Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:19.7216285Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:19.7216508Z env: 2025-12-04T09:22:19.7216762Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:19.7216932Z ##[endgroup] 2025-12-04T09:22:19.7226751Z ##[group]Run set -x 2025-12-04T09:22:19.7226937Z set -x 2025-12-04T09:22:19.7227082Z set +e 2025-12-04T09:22:19.7227228Z  2025-12-04T09:22:19.7227372Z login() { 2025-12-04T09:22:19.7227652Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-12-04T09:22:19.7227953Z } 2025-12-04T09:22:19.7228091Z  2025-12-04T09:22:19.7228256Z retry () { 2025-12-04T09:22:19.7228425Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-12-04T09:22:19.7228621Z } 2025-12-04T09:22:19.7228758Z  2025-12-04T09:22:19.7228910Z retry login "${DOCKER_REGISTRY}" 2025-12-04T09:22:19.7229088Z  2025-12-04T09:22:19.7229369Z IMAGE_SIZE=$(docker manifest inspect "${DOCKER_IMAGE}" | jq '[.layers[].size, .config.size] | add / 1024 / 1024') 2025-12-04T09:22:19.7229744Z echo "Compressed size of image in MB: ${IMAGE_SIZE}" 2025-12-04T09:22:19.7229959Z  2025-12-04T09:22:19.7230097Z set -e 2025-12-04T09:22:19.7230520Z # ignore output since only exit code is used for conditional 2025-12-04T09:22:19.7230849Z # only pull docker image if it's not available locally 2025-12-04T09:22:19.7231184Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2025-12-04T09:22:19.7231510Z  retry docker pull "${DOCKER_IMAGE}" 2025-12-04T09:22:19.7231727Z fi 2025-12-04T09:22:19.7235544Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:19.7235796Z env: 2025-12-04T09:22:19.7235965Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:19.7236505Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:19.7237123Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:19.7237381Z ##[endgroup] 2025-12-04T09:22:19.7259199Z + set +e 2025-12-04T09:22:19.7261439Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:19.7261951Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:19.7267809Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:19.7272296Z + aws ecr get-login-password --region us-east-1 2025-12-04T09:22:20.1147396Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:22:20.1147721Z Login Succeeded 2025-12-04T09:22:20.1153078Z Configure a credential helper to remove this warning. See 2025-12-04T09:22:20.1155084Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:22:20.1155428Z 2025-12-04T09:22:20.1171575Z ++ jq '[.layers[].size, .config.size] | add / 1024 / 1024' 2025-12-04T09:22:20.1176368Z ++ docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:20.3214251Z + IMAGE_SIZE=29010.187264442444 2025-12-04T09:22:20.3214552Z Compressed size of image in MB: 29010.187264442444 2025-12-04T09:22:20.3215025Z + echo 'Compressed size of image in MB: 29010.187264442444' 2025-12-04T09:22:20.3215805Z + set -e 2025-12-04T09:22:20.3216735Z + docker inspect --type=image 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:20.3327721Z + retry docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:20.3328759Z + docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:20.5752575Z pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a: Pulling from pytorch/ci-image 2025-12-04T09:22:20.5757670Z 63e5bc7682b8: Pulling fs layer 2025-12-04T09:22:20.5760317Z 6dc15eca5138: Pulling fs layer 2025-12-04T09:22:20.5760626Z 459f4df18f07: Pulling fs layer 2025-12-04T09:22:20.5760816Z 821085416919: Pulling fs layer 2025-12-04T09:22:20.5761011Z 3738646b3d92: Pulling fs layer 2025-12-04T09:22:20.5761190Z aa9a10a37b39: Pulling fs layer 2025-12-04T09:22:20.5761393Z 55bac08f3e18: Pulling fs layer 2025-12-04T09:22:20.5761582Z acf6468f6aad: Pulling fs layer 2025-12-04T09:22:20.5761924Z c0c31a0e69b7: Pulling fs layer 2025-12-04T09:22:20.5762111Z 0f7d0dc70d2d: Pulling fs layer 2025-12-04T09:22:20.5762315Z 58c4d15d3bc1: Pulling fs layer 2025-12-04T09:22:20.5762515Z 2f1cc47b61e1: Pulling fs layer 2025-12-04T09:22:20.5762712Z d2472a74103d: Pulling fs layer 2025-12-04T09:22:20.5762915Z d1fba7293688: Pulling fs layer 2025-12-04T09:22:20.5763113Z 4f4fb700ef54: Pulling fs layer 2025-12-04T09:22:20.5763290Z 6a9b03ce41a7: Pulling fs layer 2025-12-04T09:22:20.5763460Z 3519ac15be79: Pulling fs layer 2025-12-04T09:22:20.5763639Z dabb51b819a0: Pulling fs layer 2025-12-04T09:22:20.5763818Z 8f36833a24d0: Pulling fs layer 2025-12-04T09:22:20.5763988Z ab53c5a853a5: Pulling fs layer 2025-12-04T09:22:20.5764166Z 025a0e5e6ac1: Pulling fs layer 2025-12-04T09:22:20.5764345Z ce3394c8f210: Pulling fs layer 2025-12-04T09:22:20.5764515Z a5c3888c3a0c: Pulling fs layer 2025-12-04T09:22:20.5764698Z ed902a3a4e3b: Pulling fs layer 2025-12-04T09:22:20.5764881Z b4e1efca22be: Pulling fs layer 2025-12-04T09:22:20.5765057Z 79fcfd297d9f: Pulling fs layer 2025-12-04T09:22:20.5765238Z 7bd89c134b49: Pulling fs layer 2025-12-04T09:22:20.5765418Z b7d40b4fd1b9: Pulling fs layer 2025-12-04T09:22:20.5765591Z dd1cdd87320d: Pulling fs layer 2025-12-04T09:22:20.5765775Z c21a6e1cd03a: Pulling fs layer 2025-12-04T09:22:20.5765956Z 0aa3ea0b2754: Pulling fs layer 2025-12-04T09:22:20.5766140Z 74e0bdbb05d3: Pulling fs layer 2025-12-04T09:22:20.5766313Z ba3aee7dcf03: Pulling fs layer 2025-12-04T09:22:20.5766492Z 8c8d59b8759d: Pulling fs layer 2025-12-04T09:22:20.5766673Z 3c960865867b: Pulling fs layer 2025-12-04T09:22:20.5766842Z 074db79e3832: Pulling fs layer 2025-12-04T09:22:20.5767016Z 821085416919: Waiting 2025-12-04T09:22:20.5767187Z 226748f4ff23: Pulling fs layer 2025-12-04T09:22:20.5767357Z 75d817336f1d: Pulling fs layer 2025-12-04T09:22:20.5767535Z 850870a17b94: Pulling fs layer 2025-12-04T09:22:20.5767714Z e6225129924f: Pulling fs layer 2025-12-04T09:22:20.5767884Z 5cd6b9ec5c3c: Pulling fs layer 2025-12-04T09:22:20.5768064Z 613d2b1f3db8: Pulling fs layer 2025-12-04T09:22:20.5768244Z 388169fffe8a: Pulling fs layer 2025-12-04T09:22:20.5768418Z 6ca498e78b7b: Pulling fs layer 2025-12-04T09:22:20.5768596Z 952b753ec7be: Pulling fs layer 2025-12-04T09:22:20.5768768Z 3738646b3d92: Waiting 2025-12-04T09:22:20.5768923Z aa9a10a37b39: Waiting 2025-12-04T09:22:20.5769084Z 55bac08f3e18: Waiting 2025-12-04T09:22:20.5769255Z 5b6f6de4bdd1: Pulling fs layer 2025-12-04T09:22:20.5769438Z fa23d9952f3c: Pulling fs layer 2025-12-04T09:22:20.5769607Z 79fcfd297d9f: Waiting 2025-12-04T09:22:20.5769770Z acf6468f6aad: Waiting 2025-12-04T09:22:20.5769935Z c2aa9ea4d09a: Pulling fs layer 2025-12-04T09:22:20.5770102Z c0c31a0e69b7: Waiting 2025-12-04T09:22:20.5770270Z a9bbc9e426d3: Pulling fs layer 2025-12-04T09:22:20.5770443Z 7bd89c134b49: Waiting 2025-12-04T09:22:20.5770595Z b7d40b4fd1b9: Waiting 2025-12-04T09:22:20.5770751Z 0f7d0dc70d2d: Waiting 2025-12-04T09:22:20.5771156Z 4f4fb700ef54: Waiting 2025-12-04T09:22:20.5771324Z 82f2371f6dc2: Pulling fs layer 2025-12-04T09:22:20.5771502Z dd1cdd87320d: Waiting 2025-12-04T09:22:20.5771665Z 6a9b03ce41a7: Waiting 2025-12-04T09:22:20.5771817Z c21a6e1cd03a: Waiting 2025-12-04T09:22:20.5772019Z 58c4d15d3bc1: Waiting 2025-12-04T09:22:20.5772171Z 2f1cc47b61e1: Waiting 2025-12-04T09:22:20.5772329Z c87301921afa: Pulling fs layer 2025-12-04T09:22:20.5772488Z 0aa3ea0b2754: Waiting 2025-12-04T09:22:20.5772746Z 384d566a822f: Pulling fs layer 2025-12-04T09:22:20.5772920Z 32ddcdd5dd69: Pulling fs layer 2025-12-04T09:22:20.5773078Z d2472a74103d: Waiting 2025-12-04T09:22:20.5773227Z d1fba7293688: Waiting 2025-12-04T09:22:20.5773382Z a116a32ceaa1: Pulling fs layer 2025-12-04T09:22:20.5773548Z f6fbbbe6067e: Pulling fs layer 2025-12-04T09:22:20.5773713Z 3519ac15be79: Waiting 2025-12-04T09:22:20.5773864Z 648018658875: Pulling fs layer 2025-12-04T09:22:20.5774028Z 074db79e3832: Waiting 2025-12-04T09:22:20.5774173Z dabb51b819a0: Waiting 2025-12-04T09:22:20.5774327Z 8f36833a24d0: Waiting 2025-12-04T09:22:20.5774484Z 5e2f7eac20ad: Pulling fs layer 2025-12-04T09:22:20.5774644Z ba3aee7dcf03: Waiting 2025-12-04T09:22:20.5774797Z 226748f4ff23: Waiting 2025-12-04T09:22:20.5774950Z 0d1243af7593: Pulling fs layer 2025-12-04T09:22:20.5775106Z 75d817336f1d: Waiting 2025-12-04T09:22:20.5775256Z ab53c5a853a5: Waiting 2025-12-04T09:22:20.5775412Z 1f6d5b941ea1: Pulling fs layer 2025-12-04T09:22:20.5775577Z 025a0e5e6ac1: Waiting 2025-12-04T09:22:20.5775726Z ce3394c8f210: Waiting 2025-12-04T09:22:20.5775882Z 4444f70d73ce: Pulling fs layer 2025-12-04T09:22:20.5776037Z 8c8d59b8759d: Waiting 2025-12-04T09:22:20.5776186Z a5c3888c3a0c: Waiting 2025-12-04T09:22:20.5776336Z b4e1efca22be: Waiting 2025-12-04T09:22:20.5776485Z bf45550ddbad: Pulling fs layer 2025-12-04T09:22:20.5776659Z 30fff4f4bad3: Pulling fs layer 2025-12-04T09:22:20.5776833Z cca4db1ba155: Pulling fs layer 2025-12-04T09:22:20.5776990Z 850870a17b94: Waiting 2025-12-04T09:22:20.5777140Z c87301921afa: Waiting 2025-12-04T09:22:20.5777288Z e6225129924f: Waiting 2025-12-04T09:22:20.5777434Z 384d566a822f: Waiting 2025-12-04T09:22:20.5777584Z f0ffa379f4eb: Pulling fs layer 2025-12-04T09:22:20.5777758Z df645e678f95: Pulling fs layer 2025-12-04T09:22:20.5777921Z 5cd6b9ec5c3c: Waiting 2025-12-04T09:22:20.5778070Z 52ad6c17d103: Pulling fs layer 2025-12-04T09:22:20.5778234Z 74e0bdbb05d3: Waiting 2025-12-04T09:22:20.5778387Z 34408374e32c: Pulling fs layer 2025-12-04T09:22:20.5778547Z 3c960865867b: Waiting 2025-12-04T09:22:20.5778693Z 613d2b1f3db8: Waiting 2025-12-04T09:22:20.5778846Z 27803b661d9e: Pulling fs layer 2025-12-04T09:22:20.5779003Z 388169fffe8a: Waiting 2025-12-04T09:22:20.5779159Z 07dafc893dea: Pulling fs layer 2025-12-04T09:22:20.5779324Z 6ca498e78b7b: Waiting 2025-12-04T09:22:20.5779465Z 952b753ec7be: Waiting 2025-12-04T09:22:20.5779613Z 5b6f6de4bdd1: Waiting 2025-12-04T09:22:20.5779760Z 648018658875: Waiting 2025-12-04T09:22:20.5779899Z 5e2f7eac20ad: Waiting 2025-12-04T09:22:20.5780050Z f0ffa379f4eb: Waiting 2025-12-04T09:22:20.5780198Z df645e678f95: Waiting 2025-12-04T09:22:20.5780335Z 0d1243af7593: Waiting 2025-12-04T09:22:20.5780483Z 1f6d5b941ea1: Waiting 2025-12-04T09:22:20.5780630Z 52ad6c17d103: Waiting 2025-12-04T09:22:20.5780777Z 34408374e32c: Waiting 2025-12-04T09:22:20.5780915Z 4444f70d73ce: Waiting 2025-12-04T09:22:20.5781061Z 27803b661d9e: Waiting 2025-12-04T09:22:20.5781209Z bf45550ddbad: Waiting 2025-12-04T09:22:20.5781350Z 30fff4f4bad3: Waiting 2025-12-04T09:22:20.5781504Z 07dafc893dea: Waiting 2025-12-04T09:22:20.5781654Z cca4db1ba155: Waiting 2025-12-04T09:22:20.5781798Z a116a32ceaa1: Waiting 2025-12-04T09:22:20.5781950Z fa23d9952f3c: Waiting 2025-12-04T09:22:20.5782102Z f6fbbbe6067e: Waiting 2025-12-04T09:22:20.5782246Z c2aa9ea4d09a: Waiting 2025-12-04T09:22:20.5782397Z a9bbc9e426d3: Waiting 2025-12-04T09:22:20.5782549Z 82f2371f6dc2: Waiting 2025-12-04T09:22:20.5782690Z 32ddcdd5dd69: Waiting 2025-12-04T09:22:20.5782873Z ed902a3a4e3b: Waiting 2025-12-04T09:22:20.6784787Z 6dc15eca5138: Verifying Checksum 2025-12-04T09:22:20.6790000Z 6dc15eca5138: Download complete 2025-12-04T09:22:20.7597008Z 821085416919: Verifying Checksum 2025-12-04T09:22:20.7597470Z 821085416919: Download complete 2025-12-04T09:22:20.8397830Z 3738646b3d92: Verifying Checksum 2025-12-04T09:22:20.8398321Z 3738646b3d92: Download complete 2025-12-04T09:22:20.9267420Z aa9a10a37b39: Verifying Checksum 2025-12-04T09:22:20.9268067Z aa9a10a37b39: Download complete 2025-12-04T09:22:20.9327825Z 63e5bc7682b8: Download complete 2025-12-04T09:22:21.0158339Z 55bac08f3e18: Verifying Checksum 2025-12-04T09:22:21.0158615Z 55bac08f3e18: Download complete 2025-12-04T09:22:21.0220852Z acf6468f6aad: Download complete 2025-12-04T09:22:21.1164448Z 0f7d0dc70d2d: Verifying Checksum 2025-12-04T09:22:21.1164719Z 0f7d0dc70d2d: Download complete 2025-12-04T09:22:21.2066958Z 58c4d15d3bc1: Verifying Checksum 2025-12-04T09:22:21.2067404Z 58c4d15d3bc1: Download complete 2025-12-04T09:22:21.2819516Z 2f1cc47b61e1: Download complete 2025-12-04T09:22:21.3449517Z d2472a74103d: Verifying Checksum 2025-12-04T09:22:21.3455121Z d2472a74103d: Download complete 2025-12-04T09:22:22.1157444Z 63e5bc7682b8: Pull complete 2025-12-04T09:22:22.1468284Z 6dc15eca5138: Pull complete 2025-12-04T09:22:22.6325247Z c0c31a0e69b7: Verifying Checksum 2025-12-04T09:22:22.6325536Z c0c31a0e69b7: Download complete 2025-12-04T09:22:22.6428765Z 4f4fb700ef54: Verifying Checksum 2025-12-04T09:22:22.6432208Z 4f4fb700ef54: Download complete 2025-12-04T09:22:22.7290877Z 6a9b03ce41a7: Verifying Checksum 2025-12-04T09:22:22.7291387Z 6a9b03ce41a7: Download complete 2025-12-04T09:22:22.8568544Z 3519ac15be79: Verifying Checksum 2025-12-04T09:22:22.8572318Z 3519ac15be79: Download complete 2025-12-04T09:22:22.9417115Z dabb51b819a0: Verifying Checksum 2025-12-04T09:22:22.9422454Z dabb51b819a0: Download complete 2025-12-04T09:22:23.1265680Z 8f36833a24d0: Verifying Checksum 2025-12-04T09:22:23.1270158Z 8f36833a24d0: Download complete 2025-12-04T09:22:23.2049855Z ab53c5a853a5: Verifying Checksum 2025-12-04T09:22:23.2051942Z ab53c5a853a5: Download complete 2025-12-04T09:22:23.2859309Z 025a0e5e6ac1: Verifying Checksum 2025-12-04T09:22:23.2859748Z 025a0e5e6ac1: Download complete 2025-12-04T09:22:23.3473940Z ce3394c8f210: Verifying Checksum 2025-12-04T09:22:23.3475412Z ce3394c8f210: Download complete 2025-12-04T09:22:23.4429460Z a5c3888c3a0c: Verifying Checksum 2025-12-04T09:22:23.4429752Z a5c3888c3a0c: Download complete 2025-12-04T09:22:23.5353058Z ed902a3a4e3b: Verifying Checksum 2025-12-04T09:22:23.5353373Z ed902a3a4e3b: Download complete 2025-12-04T09:22:23.6098091Z b4e1efca22be: Verifying Checksum 2025-12-04T09:22:23.6100631Z b4e1efca22be: Download complete 2025-12-04T09:22:23.6888107Z 79fcfd297d9f: Verifying Checksum 2025-12-04T09:22:23.6890711Z 79fcfd297d9f: Download complete 2025-12-04T09:22:23.7733303Z 7bd89c134b49: Verifying Checksum 2025-12-04T09:22:23.7736064Z 7bd89c134b49: Download complete 2025-12-04T09:22:24.1192506Z 459f4df18f07: Download complete 2025-12-04T09:22:24.2044179Z dd1cdd87320d: Download complete 2025-12-04T09:22:24.2966797Z c21a6e1cd03a: Verifying Checksum 2025-12-04T09:22:24.2967266Z c21a6e1cd03a: Download complete 2025-12-04T09:22:24.3792447Z 0aa3ea0b2754: Verifying Checksum 2025-12-04T09:22:24.3794533Z 0aa3ea0b2754: Download complete 2025-12-04T09:22:24.4490703Z 74e0bdbb05d3: Download complete 2025-12-04T09:22:24.9421195Z ba3aee7dcf03: Verifying Checksum 2025-12-04T09:22:24.9425530Z ba3aee7dcf03: Download complete 2025-12-04T09:22:24.9997411Z 8c8d59b8759d: Verifying Checksum 2025-12-04T09:22:24.9998775Z 8c8d59b8759d: Download complete 2025-12-04T09:22:25.0907500Z 3c960865867b: Download complete 2025-12-04T09:22:25.1660802Z 074db79e3832: Verifying Checksum 2025-12-04T09:22:25.1661233Z 074db79e3832: Download complete 2025-12-04T09:22:25.2467393Z 226748f4ff23: Download complete 2025-12-04T09:22:25.3118212Z 75d817336f1d: Verifying Checksum 2025-12-04T09:22:25.3119811Z 75d817336f1d: Download complete 2025-12-04T09:22:28.8851439Z b7d40b4fd1b9: Verifying Checksum 2025-12-04T09:22:28.8852119Z b7d40b4fd1b9: Download complete 2025-12-04T09:22:28.9883753Z e6225129924f: Download complete 2025-12-04T09:22:29.0717832Z 5cd6b9ec5c3c: Verifying Checksum 2025-12-04T09:22:29.0718365Z 5cd6b9ec5c3c: Download complete 2025-12-04T09:22:29.1392447Z 613d2b1f3db8: Verifying Checksum 2025-12-04T09:22:29.1392719Z 613d2b1f3db8: Download complete 2025-12-04T09:22:33.2725946Z 388169fffe8a: Verifying Checksum 2025-12-04T09:22:33.2726245Z 388169fffe8a: Download complete 2025-12-04T09:22:36.0327604Z 459f4df18f07: Pull complete 2025-12-04T09:22:36.3029695Z 821085416919: Pull complete 2025-12-04T09:22:36.6435433Z 3738646b3d92: Pull complete 2025-12-04T09:22:36.8350641Z aa9a10a37b39: Pull complete 2025-12-04T09:22:37.0109309Z 55bac08f3e18: Pull complete 2025-12-04T09:22:37.1948926Z acf6468f6aad: Pull complete 2025-12-04T09:22:41.0679356Z c0c31a0e69b7: Pull complete 2025-12-04T09:22:41.3229927Z 0f7d0dc70d2d: Pull complete 2025-12-04T09:22:41.6167140Z 58c4d15d3bc1: Pull complete 2025-12-04T09:22:41.8713376Z 2f1cc47b61e1: Pull complete 2025-12-04T09:22:42.1106930Z d2472a74103d: Pull complete 2025-12-04T09:23:03.2816255Z d1fba7293688: Verifying Checksum 2025-12-04T09:23:03.2821187Z d1fba7293688: Download complete 2025-12-04T09:23:03.3755312Z 952b753ec7be: Download complete 2025-12-04T09:23:03.4720283Z 5b6f6de4bdd1: Verifying Checksum 2025-12-04T09:23:03.4720818Z 5b6f6de4bdd1: Download complete 2025-12-04T09:23:03.5555723Z fa23d9952f3c: Verifying Checksum 2025-12-04T09:23:03.5558032Z fa23d9952f3c: Download complete 2025-12-04T09:23:03.6387407Z c2aa9ea4d09a: Verifying Checksum 2025-12-04T09:23:03.6387778Z c2aa9ea4d09a: Download complete 2025-12-04T09:23:03.7145255Z a9bbc9e426d3: Verifying Checksum 2025-12-04T09:23:03.7145682Z a9bbc9e426d3: Download complete 2025-12-04T09:23:03.8351854Z 82f2371f6dc2: Verifying Checksum 2025-12-04T09:23:03.8352310Z 82f2371f6dc2: Download complete 2025-12-04T09:23:03.9275232Z c87301921afa: Verifying Checksum 2025-12-04T09:23:03.9277388Z c87301921afa: Download complete 2025-12-04T09:23:04.0273322Z 384d566a822f: Verifying Checksum 2025-12-04T09:23:04.0277965Z 384d566a822f: Download complete 2025-12-04T09:23:04.1153528Z 32ddcdd5dd69: Verifying Checksum 2025-12-04T09:23:04.1158072Z 32ddcdd5dd69: Download complete 2025-12-04T09:23:04.2149332Z a116a32ceaa1: Verifying Checksum 2025-12-04T09:23:04.2149789Z a116a32ceaa1: Download complete 2025-12-04T09:23:04.2983092Z f6fbbbe6067e: Verifying Checksum 2025-12-04T09:23:04.2989921Z f6fbbbe6067e: Download complete 2025-12-04T09:23:04.3956588Z 648018658875: Verifying Checksum 2025-12-04T09:23:04.3956889Z 648018658875: Download complete 2025-12-04T09:23:04.4728909Z 5e2f7eac20ad: Download complete 2025-12-04T09:23:04.6012213Z 0d1243af7593: Verifying Checksum 2025-12-04T09:23:04.6012638Z 0d1243af7593: Download complete 2025-12-04T09:23:04.6960853Z 1f6d5b941ea1: Verifying Checksum 2025-12-04T09:23:04.6961172Z 1f6d5b941ea1: Download complete 2025-12-04T09:23:04.7766853Z 4444f70d73ce: Download complete 2025-12-04T09:23:04.8641286Z bf45550ddbad: Verifying Checksum 2025-12-04T09:23:04.8641652Z bf45550ddbad: Download complete 2025-12-04T09:23:08.2724024Z 30fff4f4bad3: Verifying Checksum 2025-12-04T09:23:08.2724462Z 30fff4f4bad3: Download complete 2025-12-04T09:23:08.3656756Z cca4db1ba155: Verifying Checksum 2025-12-04T09:23:08.3657207Z cca4db1ba155: Download complete 2025-12-04T09:23:08.4538325Z f0ffa379f4eb: Verifying Checksum 2025-12-04T09:23:08.4542631Z f0ffa379f4eb: Download complete 2025-12-04T09:23:08.5451598Z df645e678f95: Verifying Checksum 2025-12-04T09:23:08.5452071Z df645e678f95: Download complete 2025-12-04T09:23:08.6250661Z 52ad6c17d103: Verifying Checksum 2025-12-04T09:23:08.6250985Z 52ad6c17d103: Download complete 2025-12-04T09:23:08.7219627Z 34408374e32c: Verifying Checksum 2025-12-04T09:23:08.7220049Z 34408374e32c: Download complete 2025-12-04T09:23:08.8224677Z 27803b661d9e: Verifying Checksum 2025-12-04T09:23:08.8225232Z 27803b661d9e: Download complete 2025-12-04T09:23:09.5393092Z 07dafc893dea: Verifying Checksum 2025-12-04T09:23:09.5394885Z 07dafc893dea: Download complete 2025-12-04T09:23:58.5980287Z 6ca498e78b7b: Verifying Checksum 2025-12-04T09:23:58.5980656Z 6ca498e78b7b: Download complete 2025-12-04T09:24:26.8472882Z d1fba7293688: Pull complete 2025-12-04T09:24:27.2547135Z 4f4fb700ef54: Pull complete 2025-12-04T09:24:27.6368622Z 6a9b03ce41a7: Pull complete 2025-12-04T09:24:28.1203255Z 3519ac15be79: Pull complete 2025-12-04T09:24:28.5577843Z dabb51b819a0: Pull complete 2025-12-04T09:24:29.0491494Z 8f36833a24d0: Pull complete 2025-12-04T09:24:29.4776670Z ab53c5a853a5: Pull complete 2025-12-04T09:24:29.7007324Z 025a0e5e6ac1: Pull complete 2025-12-04T09:24:29.9119354Z ce3394c8f210: Pull complete 2025-12-04T09:24:30.2609333Z a5c3888c3a0c: Pull complete 2025-12-04T09:24:30.7985796Z ed902a3a4e3b: Pull complete 2025-12-04T09:24:31.2015293Z b4e1efca22be: Pull complete 2025-12-04T09:24:31.8570933Z 79fcfd297d9f: Pull complete 2025-12-04T09:24:32.3938267Z 7bd89c134b49: Pull complete 2025-12-04T09:24:45.1773958Z b7d40b4fd1b9: Pull complete 2025-12-04T09:24:45.6045332Z dd1cdd87320d: Pull complete 2025-12-04T09:24:46.0717560Z c21a6e1cd03a: Pull complete 2025-12-04T09:24:46.7207474Z 0aa3ea0b2754: Pull complete 2025-12-04T09:24:47.2192541Z 74e0bdbb05d3: Pull complete 2025-12-04T09:24:47.9691723Z ba3aee7dcf03: Pull complete 2025-12-04T09:24:48.2552428Z 8c8d59b8759d: Pull complete 2025-12-04T09:24:48.6220925Z 3c960865867b: Pull complete 2025-12-04T09:24:49.4090206Z 074db79e3832: Pull complete 2025-12-04T09:24:49.7497534Z 226748f4ff23: Pull complete 2025-12-04T09:24:50.1794058Z 75d817336f1d: Pull complete 2025-12-04T09:26:23.5468661Z 850870a17b94: Verifying Checksum 2025-12-04T09:26:23.5472648Z 850870a17b94: Download complete 2025-12-04T09:31:05.1737926Z 850870a17b94: Pull complete 2025-12-04T09:31:05.6344398Z e6225129924f: Pull complete 2025-12-04T09:31:06.1319279Z 5cd6b9ec5c3c: Pull complete 2025-12-04T09:31:07.1084069Z 613d2b1f3db8: Pull complete 2025-12-04T09:31:10.5922800Z 388169fffe8a: Pull complete 2025-12-04T09:33:55.5758576Z 6ca498e78b7b: Pull complete 2025-12-04T09:33:55.9951157Z 952b753ec7be: Pull complete 2025-12-04T09:33:56.4483152Z 5b6f6de4bdd1: Pull complete 2025-12-04T09:33:57.4214624Z fa23d9952f3c: Pull complete 2025-12-04T09:33:57.9177843Z c2aa9ea4d09a: Pull complete 2025-12-04T09:33:58.3374420Z a9bbc9e426d3: Pull complete 2025-12-04T09:33:59.0880270Z 82f2371f6dc2: Pull complete 2025-12-04T09:33:59.5305837Z c87301921afa: Pull complete 2025-12-04T09:33:59.5577447Z 384d566a822f: Pull complete 2025-12-04T09:33:59.6156160Z 32ddcdd5dd69: Pull complete 2025-12-04T09:33:59.6726396Z a116a32ceaa1: Pull complete 2025-12-04T09:33:59.6991040Z f6fbbbe6067e: Pull complete 2025-12-04T09:33:59.7508167Z 648018658875: Pull complete 2025-12-04T09:33:59.7767848Z 5e2f7eac20ad: Pull complete 2025-12-04T09:33:59.8308071Z 0d1243af7593: Pull complete 2025-12-04T09:33:59.8590939Z 1f6d5b941ea1: Pull complete 2025-12-04T09:33:59.9128208Z 4444f70d73ce: Pull complete 2025-12-04T09:33:59.9415403Z bf45550ddbad: Pull complete 2025-12-04T09:34:10.0312250Z 30fff4f4bad3: Pull complete 2025-12-04T09:34:10.4200093Z cca4db1ba155: Pull complete 2025-12-04T09:34:10.7344036Z f0ffa379f4eb: Pull complete 2025-12-04T09:34:11.2260894Z df645e678f95: Pull complete 2025-12-04T09:34:11.7373215Z 52ad6c17d103: Pull complete 2025-12-04T09:34:12.1970207Z 34408374e32c: Pull complete 2025-12-04T09:34:13.2190370Z 27803b661d9e: Pull complete 2025-12-04T09:34:16.4783783Z 07dafc893dea: Pull complete 2025-12-04T09:34:17.0286816Z Digest: sha256:b178ee928adfcab963ebbc6ea05ea1a0f0c605bb095e9076f010bf92d150869d 2025-12-04T09:34:17.1432974Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:34:17.1836806Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:34:17.1886864Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:34:17.1887468Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:34:17.1895515Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:17.1895761Z env: 2025-12-04T09:34:17.1895915Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:17.1896094Z ##[endgroup] 2025-12-04T09:34:17.2022764Z ##[group]Run pytorch/test-infra/.github/actions/setup-nvidia@main 2025-12-04T09:34:17.2023238Z with: 2025-12-04T09:34:17.2023414Z driver-version: 580.82.07 2025-12-04T09:34:17.2023596Z env: 2025-12-04T09:34:17.2023757Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:17.2023944Z ##[endgroup] 2025-12-04T09:34:17.2068881Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:34:17.2069394Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:34:17.2073542Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:17.2073781Z env: 2025-12-04T09:34:17.2073934Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:17.2074098Z ##[endgroup] 2025-12-04T09:34:17.2122624Z ##[group]Run set -euo pipefail 2025-12-04T09:34:17.2122883Z set -euo pipefail 2025-12-04T09:34:17.2123118Z  2025-12-04T09:34:17.2123288Z has_gpu=false 2025-12-04T09:34:17.2123491Z devices="" 2025-12-04T09:34:17.2123672Z  2025-12-04T09:34:17.2123882Z if command -v nvidia-smi >/dev/null 2>&1; then 2025-12-04T09:34:17.2124179Z  if nvidia-smi -L >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:34:17.2124418Z  has_gpu=true 2025-12-04T09:34:17.2124616Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:34:17.2124810Z  fi 2025-12-04T09:34:17.2124959Z fi 2025-12-04T09:34:17.2125104Z  2025-12-04T09:34:17.2125253Z if [ "$has_gpu" = false ]; then 2025-12-04T09:34:17.2125499Z  if ls /dev/nvidia* >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:34:17.2125737Z  has_gpu=true 2025-12-04T09:34:17.2125934Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:34:17.2126131Z  fi 2025-12-04T09:34:17.2126282Z fi 2025-12-04T09:34:17.2126427Z  2025-12-04T09:34:17.2126631Z if [ "$has_gpu" = false ] && command -v lspci >/dev/null 2>&1; then 2025-12-04T09:34:17.2126946Z  if lspci | grep -i 'nvidia' >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:34:17.2127200Z  has_gpu=true 2025-12-04T09:34:17.2127389Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:34:17.2127600Z  fi 2025-12-04T09:34:17.2127743Z fi 2025-12-04T09:34:17.2127876Z  2025-12-04T09:34:17.2128076Z printf 'HAS_NVIDIA=%s\n' "$has_gpu" >> "$GITHUB_OUTPUT" 2025-12-04T09:34:17.2128395Z printf 'DETECTED_DEVICES<> "$GITHUB_OUTPUT" 2025-12-04T09:34:17.2132376Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:17.2132610Z env: 2025-12-04T09:34:17.2132768Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:17.2132945Z ##[endgroup] 2025-12-04T09:34:17.2470648Z ##[group]Run if [ "${HAS_NVIDIA}" = "true" ]; then 2025-12-04T09:34:17.2470921Z if [ "${HAS_NVIDIA}" = "true" ]; then 2025-12-04T09:34:17.2471237Z  echo "HAS_NVIDIA_GPU=true" >> "${GITHUB_ENV}" 2025-12-04T09:34:17.2471591Z  echo "GPU_FLAG=--gpus all -e NVIDIA_DRIVER_CAPABILITIES=all" >> "${GITHUB_ENV}" 2025-12-04T09:34:17.2471862Z else 2025-12-04T09:34:17.2472086Z  echo "HAS_NVIDIA_GPU=false" >> "${GITHUB_ENV}" 2025-12-04T09:34:17.2472288Z fi 2025-12-04T09:34:17.2476067Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:17.2476302Z env: 2025-12-04T09:34:17.2476449Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:17.2476622Z HAS_NVIDIA: false 2025-12-04T09:34:17.2476777Z ##[endgroup] 2025-12-04T09:34:17.2540977Z Prepare all required actions 2025-12-04T09:34:17.2558340Z ##[group]Run ./.github/actions/get-workflow-job-id 2025-12-04T09:34:17.2558561Z with: 2025-12-04T09:34:17.2559102Z github-token: *** 2025-12-04T09:34:17.2559276Z env: 2025-12-04T09:34:17.2559430Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:17.2559626Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:17.2560790Z ##[endgroup] 2025-12-04T09:34:17.2570868Z ##[group]Run set -eux 2025-12-04T09:34:17.2571058Z set -eux 2025-12-04T09:34:17.2571361Z python3 .github/scripts/get_workflow_job_id.py "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-12-04T09:34:17.2575176Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:17.2575410Z env: 2025-12-04T09:34:17.2575558Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:17.2575725Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:17.2576045Z GITHUB_TOKEN: *** 2025-12-04T09:34:17.2576202Z ##[endgroup] 2025-12-04T09:34:17.2597364Z + python3 .github/scripts/get_workflow_job_id.py 19923066595 i-0c3ca63174fa64465 2025-12-04T09:34:19.5666584Z Setting output job-id=57118563344 2025-12-04T09:34:19.5667348Z Setting output job-name=periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx) 2025-12-04T09:34:19.5767776Z ##[group]Run python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-12-04T09:34:19.5768242Z python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-12-04T09:34:19.5768796Z python3 -m tools.stats.monitor --log-interval "$MONITOR_LOG_INTERVAL" --data-collect-interval "$MONITOR_DATA_COLLECT_INTERVAL" > usage_log.txt 2>&1 & 2025-12-04T09:34:19.5769286Z echo "monitor-script-pid=${!}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:34:19.5774088Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:19.5774327Z env: 2025-12-04T09:34:19.5774483Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:19.5774668Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:19.5774830Z JOB_ID: 57118563344 2025-12-04T09:34:19.5775190Z JOB_NAME: periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx) 2025-12-04T09:34:19.5775609Z WORKFLOW_NAME: inductor-periodic 2025-12-04T09:34:19.5775811Z WORKFLOW_RUN_ID: 19923066595 2025-12-04T09:34:19.5776003Z MONITOR_LOG_INTERVAL: 5 2025-12-04T09:34:19.5776184Z MONITOR_DATA_COLLECT_INTERVAL: 1 2025-12-04T09:34:19.5776363Z ##[endgroup] 2025-12-04T09:34:19.8287779Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T09:34:20.0729032Z Collecting psutil==5.9.8 2025-12-04T09:34:20.0877997Z Downloading psutil-5.9.8-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (288 kB) 2025-12-04T09:34:20.1444931Z Collecting dataclasses_json==0.6.7 2025-12-04T09:34:20.1478543Z Downloading dataclasses_json-0.6.7-py3-none-any.whl (28 kB) 2025-12-04T09:34:20.1711715Z Collecting nvidia-ml-py==11.525.84 2025-12-04T09:34:20.1742527Z Downloading nvidia_ml_py-11.525.84-py3-none-any.whl (34 kB) 2025-12-04T09:34:20.2010731Z Collecting typing-inspect<1,>=0.4.0 2025-12-04T09:34:20.2041807Z Downloading typing_inspect-0.9.0-py3-none-any.whl (8.8 kB) 2025-12-04T09:34:20.2786919Z Collecting marshmallow<4.0.0,>=3.18.0 2025-12-04T09:34:20.2821609Z Downloading marshmallow-3.26.1-py3-none-any.whl (50 kB) 2025-12-04T09:34:20.3246267Z Collecting packaging>=17.0 2025-12-04T09:34:20.3281647Z Downloading packaging-25.0-py3-none-any.whl (66 kB) 2025-12-04T09:34:20.3678133Z Collecting typing-extensions>=3.7.4 2025-12-04T09:34:20.3714762Z Downloading typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-12-04T09:34:20.3898931Z Collecting mypy-extensions>=0.3.0 2025-12-04T09:34:20.3927192Z Downloading mypy_extensions-1.1.0-py3-none-any.whl (5.0 kB) 2025-12-04T09:34:20.4731075Z Installing collected packages: typing-extensions, packaging, mypy-extensions, typing-inspect, marshmallow, psutil, nvidia-ml-py, dataclasses-json 2025-12-04T09:34:20.7023685Z Successfully installed dataclasses-json-0.6.7 marshmallow-3.26.1 mypy-extensions-1.1.0 nvidia-ml-py-11.525.84 packaging-25.0 psutil-5.9.8 typing-extensions-4.15.0 typing-inspect-0.9.0 2025-12-04T09:34:20.8512881Z Prepare all required actions 2025-12-04T09:34:20.8513157Z Getting action download info 2025-12-04T09:34:21.0227086Z Download action repository 'seemethere/download-artifact-s3@v4' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-12-04T09:34:21.2972694Z Download action repository 'actions/download-artifact@v4' (SHA:d3f86a106a0bac45b974a628896c90dbdf5c8093) 2025-12-04T09:34:21.6650968Z ##[group]Run ./.github/actions/download-build-artifacts 2025-12-04T09:34:21.6651214Z with: 2025-12-04T09:34:21.6651392Z name: linux-jammy-py3.10-gcc11-build 2025-12-04T09:34:21.6651607Z s3-bucket: gha-artifacts 2025-12-04T09:34:21.6651778Z env: 2025-12-04T09:34:21.6651930Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:21.6652107Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:21.6652271Z ##[endgroup] 2025-12-04T09:34:21.6670951Z ##[group]Run seemethere/download-artifact-s3@v4 2025-12-04T09:34:21.6671168Z with: 2025-12-04T09:34:21.6671330Z name: linux-jammy-py3.10-gcc11-build 2025-12-04T09:34:21.6671567Z s3-bucket: gha-artifacts 2025-12-04T09:34:21.6671737Z region: us-east-1 2025-12-04T09:34:21.6671887Z env: 2025-12-04T09:34:21.6672044Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:21.6672208Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:21.6672374Z ##[endgroup] 2025-12-04T09:34:22.0504340Z (node:52421) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-12-04T09:34:22.0505554Z 2025-12-04T09:34:22.0506073Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-12-04T09:34:22.0506496Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-12-04T09:34:22.0506949Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-12-04T09:34:22.4070618Z Found 1 objects with prefix pytorch/pytorch/19923066595/linux-jammy-py3.10-gcc11-build/ 2025-12-04T09:34:22.4071515Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-12-04T09:34:27.1427750Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-12-04T09:34:27.1450454Z Artifact download has finished successfully 2025-12-04T09:34:27.1600043Z ##[group]Run unzip -o artifacts.zip 2025-12-04T09:34:27.1600286Z unzip -o artifacts.zip 2025-12-04T09:34:27.1604948Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:27.1605199Z env: 2025-12-04T09:34:27.1605359Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:27.1605545Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:27.1605732Z ##[endgroup] 2025-12-04T09:34:27.1677786Z Archive: artifacts.zip 2025-12-04T09:34:27.1678059Z creating: dist/ 2025-12-04T09:34:28.2217336Z inflating: dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl 2025-12-04T09:34:28.2218941Z creating: dist/vision/ 2025-12-04T09:34:28.2289518Z inflating: dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:34:28.2293991Z creating: dist/audio/ 2025-12-04T09:34:28.2313616Z inflating: dist/audio/torchaudio-2.10.0a0+e90a398-cp310-cp310-linux_x86_64.whl 2025-12-04T09:34:28.2315129Z creating: dist/ao/ 2025-12-04T09:34:28.2349762Z inflating: dist/ao/torchao-0.7.0+git51c87b6e-py3-none-any.whl 2025-12-04T09:34:28.2454373Z inflating: dist/.ninja_log 2025-12-04T09:34:28.2458520Z creating: build/custom_test_artifacts/ 2025-12-04T09:34:28.2461361Z creating: build/custom_test_artifacts/custom-op-build/ 2025-12-04T09:34:28.2461800Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/ 2025-12-04T09:34:28.2466646Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:34:28.2469880Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:34:28.2470320Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/ 2025-12-04T09:34:28.2470702Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:34:28.2471466Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:34:28.2471870Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:34:28.2472520Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:34:28.2472969Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:34:28.2473431Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:34:28.2473842Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:34:28.2474239Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:34:28.2474689Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:34:28.2475157Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:34:28.2475581Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:34:28.2476032Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:34:28.2476570Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:34:28.2477039Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:34:28.2477418Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/cmake.check_cache 2025-12-04T09:34:28.2477797Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/ 2025-12-04T09:34:28.2478215Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.ts 2025-12-04T09:34:28.2478688Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.make 2025-12-04T09:34:28.2479141Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/depend.make 2025-12-04T09:34:28.2479559Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.txt 2025-12-04T09:34:28.2479989Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/cmake_clean.cmake 2025-12-04T09:34:28.2480428Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/build.make 2025-12-04T09:34:28.2480862Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/DependInfo.cmake 2025-12-04T09:34:28.2481290Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/flags.make 2025-12-04T09:34:28.2481719Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/progress.make 2025-12-04T09:34:28.2491902Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o.d 2025-12-04T09:34:28.2657764Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o 2025-12-04T09:34:28.2659782Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/ 2025-12-04T09:34:28.2664330Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.ts 2025-12-04T09:34:28.2666152Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.make 2025-12-04T09:34:28.2666786Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/depend.make 2025-12-04T09:34:28.2672367Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.txt 2025-12-04T09:34:28.2677093Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/cmake_clean.cmake 2025-12-04T09:34:28.2678017Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/build.make 2025-12-04T09:34:28.2678658Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/DependInfo.cmake 2025-12-04T09:34:28.2679244Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/flags.make 2025-12-04T09:34:28.2679713Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/progress.make 2025-12-04T09:34:28.2680216Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o.d 2025-12-04T09:34:28.2747788Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o 2025-12-04T09:34:28.2748483Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:34:28.2753547Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:34:28.2758014Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/progress.marks 2025-12-04T09:34:28.2759855Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile2 2025-12-04T09:34:28.2760269Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile.cmake 2025-12-04T09:34:28.2760640Z inflating: build/custom_test_artifacts/custom-op-build/CMakeCache.txt 2025-12-04T09:34:28.2760975Z inflating: build/custom_test_artifacts/custom-op-build/Makefile 2025-12-04T09:34:28.2761314Z inflating: build/custom_test_artifacts/custom-op-build/cmake_install.cmake 2025-12-04T09:34:28.2896718Z inflating: build/custom_test_artifacts/custom-op-build/libcustom_ops.so 2025-12-04T09:34:28.2945541Z inflating: build/custom_test_artifacts/custom-op-build/test_custom_ops 2025-12-04T09:34:28.2947882Z creating: build/custom_test_artifacts/jit-hook-build/ 2025-12-04T09:34:28.2948416Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/ 2025-12-04T09:34:28.2948845Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:34:28.2953689Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:34:28.2957877Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/ 2025-12-04T09:34:28.2958318Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:34:28.2958743Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:34:28.2959150Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:34:28.2959611Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:34:28.2960062Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:34:28.2960515Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:34:28.2960939Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:34:28.2961357Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:34:28.2961831Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:34:28.2962483Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:34:28.2962950Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:34:28.2963443Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:34:28.2963956Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:34:28.2964784Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:34:28.2965169Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/cmake.check_cache 2025-12-04T09:34:28.2965624Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/ 2025-12-04T09:34:28.2966094Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.ts 2025-12-04T09:34:28.2966552Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.make 2025-12-04T09:34:28.2966998Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/depend.make 2025-12-04T09:34:28.2967417Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.txt 2025-12-04T09:34:28.2967874Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/cmake_clean.cmake 2025-12-04T09:34:28.2968311Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/build.make 2025-12-04T09:34:28.2968738Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/DependInfo.cmake 2025-12-04T09:34:28.2969161Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/flags.make 2025-12-04T09:34:28.2969585Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/progress.make 2025-12-04T09:34:28.2979040Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o.d 2025-12-04T09:34:28.3031798Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o 2025-12-04T09:34:28.3034813Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:34:28.3037458Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:34:28.3037930Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/progress.marks 2025-12-04T09:34:28.3038320Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile2 2025-12-04T09:34:28.3038702Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile.cmake 2025-12-04T09:34:28.3039068Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeCache.txt 2025-12-04T09:34:28.3039393Z inflating: build/custom_test_artifacts/jit-hook-build/Makefile 2025-12-04T09:34:28.3039733Z inflating: build/custom_test_artifacts/jit-hook-build/cmake_install.cmake 2025-12-04T09:34:28.3070999Z inflating: build/custom_test_artifacts/jit-hook-build/test_jit_hooks 2025-12-04T09:34:28.3076470Z creating: build/custom_test_artifacts/custom-backend-build/ 2025-12-04T09:34:28.3078147Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/ 2025-12-04T09:34:28.3078543Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:34:28.3080968Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:34:28.3081416Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/ 2025-12-04T09:34:28.3081828Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:34:28.3082338Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:34:28.3082841Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:34:28.3083351Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:34:28.3083860Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:34:28.3084295Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:34:28.3084929Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:34:28.3085434Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:34:28.3085929Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:34:28.3086432Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:34:28.3086893Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:34:28.3087366Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:34:28.3087873Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:34:28.3088320Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:34:28.3088707Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/cmake.check_cache 2025-12-04T09:34:28.3089098Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/ 2025-12-04T09:34:28.3089547Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.ts 2025-12-04T09:34:28.3090046Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.make 2025-12-04T09:34:28.3090524Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/depend.make 2025-12-04T09:34:28.3090967Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.txt 2025-12-04T09:34:28.3091428Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/cmake_clean.cmake 2025-12-04T09:34:28.3091910Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/build.make 2025-12-04T09:34:28.3092443Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/DependInfo.cmake 2025-12-04T09:34:28.3093203Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/flags.make 2025-12-04T09:34:28.3093762Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/progress.make 2025-12-04T09:34:28.3094291Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o.d 2025-12-04T09:34:28.3189542Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o 2025-12-04T09:34:28.3191488Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/ 2025-12-04T09:34:28.3192296Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.ts 2025-12-04T09:34:28.3194755Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.make 2025-12-04T09:34:28.3195405Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/depend.make 2025-12-04T09:34:28.3196033Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.txt 2025-12-04T09:34:28.3196749Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/cmake_clean.cmake 2025-12-04T09:34:28.3198870Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/build.make 2025-12-04T09:34:28.3199531Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/DependInfo.cmake 2025-12-04T09:34:28.3200098Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/flags.make 2025-12-04T09:34:28.3200790Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/progress.make 2025-12-04T09:34:28.3209936Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o.d 2025-12-04T09:34:28.3256959Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o 2025-12-04T09:34:28.3259958Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:34:28.3260827Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:34:28.3265364Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/progress.marks 2025-12-04T09:34:28.3267643Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile2 2025-12-04T09:34:28.3273649Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile.cmake 2025-12-04T09:34:28.3277689Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeCache.txt 2025-12-04T09:34:28.3279534Z inflating: build/custom_test_artifacts/custom-backend-build/Makefile 2025-12-04T09:34:28.3279986Z inflating: build/custom_test_artifacts/custom-backend-build/cmake_install.cmake 2025-12-04T09:34:28.3346404Z inflating: build/custom_test_artifacts/custom-backend-build/libcustom_backend.so 2025-12-04T09:34:28.3379105Z inflating: build/custom_test_artifacts/custom-backend-build/test_custom_backend 2025-12-04T09:34:28.3384028Z creating: build/lib/ 2025-12-04T09:34:28.3450486Z inflating: build/lib/libprotobuf-lite.a 2025-12-04T09:34:28.3829549Z inflating: build/lib/libprotobuf.a 2025-12-04T09:34:28.4257585Z inflating: build/lib/libprotoc.a 2025-12-04T09:34:28.4265703Z inflating: build/lib/libpthreadpool.a 2025-12-04T09:34:28.4271070Z inflating: build/lib/libcpuinfo.a 2025-12-04T09:34:28.4279227Z inflating: build/lib/libcpuinfo_internals.a 2025-12-04T09:34:28.4279721Z inflating: build/lib/libclog.a 2025-12-04T09:34:28.4296232Z inflating: build/lib/libpytorch_qnnpack.a 2025-12-04T09:34:28.4300706Z inflating: build/lib/libnnpack_reference_layers.a 2025-12-04T09:34:28.4458863Z inflating: build/lib/libmicrokernels-prod.a 2025-12-04T09:34:28.4473798Z inflating: build/lib/libnnpack.a 2025-12-04T09:34:28.5232931Z inflating: build/lib/libmicrokernels-all.a 2025-12-04T09:34:28.5297511Z inflating: build/lib/libgtest.a 2025-12-04T09:34:28.5308791Z inflating: build/lib/libgmock.a 2025-12-04T09:34:28.5313437Z inflating: build/lib/libgtest_main.a 2025-12-04T09:34:28.5314085Z inflating: build/lib/libgmock_main.a 2025-12-04T09:34:28.5385483Z inflating: build/lib/libXNNPACK.a 2025-12-04T09:34:28.5452782Z inflating: build/lib/libbenchmark.a 2025-12-04T09:34:28.5455003Z inflating: build/lib/libbenchmark_main.a 2025-12-04T09:34:28.5455618Z inflating: build/lib/libjitprofiling.a 2025-12-04T09:34:28.5460671Z inflating: build/lib/libittnotify.a 2025-12-04T09:34:28.5517333Z inflating: build/lib/libasmjit.a 2025-12-04T09:34:28.6509398Z inflating: build/lib/libfbgemm.a 2025-12-04T09:34:28.6535018Z inflating: build/lib/libtensorpipe_uv.a 2025-12-04T09:34:28.7000742Z inflating: build/lib/libtensorpipe.a 2025-12-04T09:34:28.7106191Z inflating: build/lib/libgloo.a 2025-12-04T09:34:28.7146060Z inflating: build/lib/libonnx_proto.a 2025-12-04T09:34:28.7758240Z inflating: build/lib/libonnx.a 2025-12-04T09:34:29.6394344Z inflating: build/lib/libdnnl.a 2025-12-04T09:34:29.6415209Z inflating: build/lib/libfmt.a 2025-12-04T09:34:29.6654054Z inflating: build/lib/libkineto.a 2025-12-04T09:34:29.6755007Z inflating: build/lib/libc10.so 2025-12-04T09:34:29.6756830Z inflating: build/lib/libtorch_global_deps.so 2025-12-04T09:34:32.3362652Z inflating: build/lib/libtorch_cpu.so 2025-12-04T09:34:32.3363499Z inflating: build/lib/libtorch.so 2025-12-04T09:34:32.3426879Z inflating: build/lib/libtorchbind_test.so 2025-12-04T09:34:32.3442908Z inflating: build/lib/libjitbackend_test.so 2025-12-04T09:34:32.3462346Z inflating: build/lib/libbackend_with_compiler.so 2025-12-04T09:34:32.3487476Z inflating: build/lib/libaoti_custom_ops.so 2025-12-04T09:34:32.3492059Z inflating: build/lib/libshm.so 2025-12-04T09:34:32.5476930Z inflating: build/lib/libtorch_python.so 2025-12-04T09:34:32.5508704Z inflating: build/lib/libnnapi_backend.so 2025-12-04T09:34:32.5510581Z creating: build/bin/ 2025-12-04T09:34:32.5511130Z creating: build/bin/CMakeFiles/ 2025-12-04T09:34:32.5511430Z inflating: build/bin/cmake_install.cmake 2025-12-04T09:34:32.5511715Z inflating: build/bin/CTestTestfile.cmake 2025-12-04T09:34:32.5909155Z inflating: build/bin/protoc-3.13.0.0 2025-12-04T09:34:32.6305886Z inflating: build/bin/protoc 2025-12-04T09:34:32.6360503Z inflating: build/bin/c10_AllocatorConfig_test 2025-12-04T09:34:32.6408239Z inflating: build/bin/c10_CompileTimeFunctionPointer_test 2025-12-04T09:34:32.6459979Z inflating: build/bin/c10_DeviceGuard_test 2025-12-04T09:34:32.6510438Z inflating: build/bin/c10_Device_test 2025-12-04T09:34:32.6567774Z inflating: build/bin/c10_DispatchKeySet_test 2025-12-04T09:34:32.6620747Z inflating: build/bin/c10_Scalar_test 2025-12-04T09:34:32.6667378Z inflating: build/bin/c10_StreamGuard_test 2025-12-04T09:34:32.6721714Z inflating: build/bin/c10_SymInt_test 2025-12-04T09:34:32.6774790Z inflating: build/bin/c10_InlineDeviceGuard_test 2025-12-04T09:34:32.6827884Z inflating: build/bin/c10_InlineStreamGuard_test 2025-12-04T09:34:32.6881638Z inflating: build/bin/c10_SizesAndStrides_test 2025-12-04T09:34:32.6948001Z inflating: build/bin/c10_cow_test 2025-12-04T09:34:32.7000963Z inflating: build/bin/c10_Bitset_test 2025-12-04T09:34:32.7048684Z inflating: build/bin/c10_ArrayRef_test 2025-12-04T09:34:32.7096868Z inflating: build/bin/c10_ConstexprCrc_test 2025-12-04T09:34:32.7147552Z inflating: build/bin/c10_DeadlockDetection_test 2025-12-04T09:34:32.7201765Z inflating: build/bin/c10_Enumerate_test 2025-12-04T09:34:32.7255028Z inflating: build/bin/c10_LeftRight_test 2025-12-04T09:34:32.7307134Z inflating: build/bin/c10_NetworkFlow_test 2025-12-04T09:34:32.7355718Z inflating: build/bin/c10_Half_test 2025-12-04T09:34:32.7408433Z inflating: build/bin/c10_IntrusiveList_test 2025-12-04T09:34:32.7456435Z inflating: build/bin/c10_Synchronized_test 2025-12-04T09:34:32.7508547Z inflating: build/bin/c10_ThreadLocal_test 2025-12-04T09:34:32.7557360Z inflating: build/bin/c10_Semaphore_test 2025-12-04T09:34:32.7611500Z inflating: build/bin/c10_bfloat16_test 2025-12-04T09:34:32.7663667Z inflating: build/bin/c10_accumulate_test 2025-12-04T09:34:32.7712140Z inflating: build/bin/c10_TypeIndex_test 2025-12-04T09:34:32.7762718Z inflating: build/bin/c10_exception_test 2025-12-04T09:34:32.7810220Z inflating: build/bin/c10_bit_cast_test 2025-12-04T09:34:32.7863787Z inflating: build/bin/c10_complex_math_test 2025-12-04T09:34:32.7912713Z inflating: build/bin/c10_error_test 2025-12-04T09:34:32.7963453Z inflating: build/bin/c10_flags_test 2025-12-04T09:34:32.8018299Z inflating: build/bin/c10_complex_test 2025-12-04T09:34:32.8065951Z inflating: build/bin/c10_irange_test 2025-12-04T09:34:32.8115336Z inflating: build/bin/c10_generic_math_test 2025-12-04T09:34:32.8164420Z inflating: build/bin/c10_nofatal_test 2025-12-04T09:34:32.8218098Z inflating: build/bin/c10_lazy_test 2025-12-04T09:34:32.8278134Z inflating: build/bin/c10_logging_test 2025-12-04T09:34:32.8418971Z inflating: build/bin/c10_intrusive_ptr_test 2025-12-04T09:34:32.8491375Z inflating: build/bin/c10_optional_test 2025-12-04T09:34:32.8543637Z inflating: build/bin/c10_registry_test 2025-12-04T09:34:32.8602122Z inflating: build/bin/c10_ordered_preserving_dict_test 2025-12-04T09:34:32.8652622Z inflating: build/bin/c10_ssize_test 2025-12-04T09:34:32.8789695Z inflating: build/bin/c10_small_vector_test 2025-12-04T09:34:32.8844710Z inflating: build/bin/c10_string_util_test 2025-12-04T09:34:32.8892560Z inflating: build/bin/c10_string_view_test 2025-12-04T09:34:32.8939913Z inflating: build/bin/c10_tempfile_test 2025-12-04T09:34:32.8984766Z inflating: build/bin/c10_intrusive_ptr_benchmark 2025-12-04T09:34:32.9037874Z inflating: build/bin/c10_typeid_test 2025-12-04T09:34:32.9555783Z inflating: build/bin/vec_test_all_types_DEFAULT 2025-12-04T09:34:33.0090056Z inflating: build/bin/vec_test_all_types_AVX512 2025-12-04T09:34:33.0631114Z inflating: build/bin/vec_test_all_types_AVX2 2025-12-04T09:34:33.0724558Z inflating: build/bin/test_aoti_abi_check 2025-12-04T09:34:33.0772257Z inflating: build/bin/test_vec_half_DEFAULT 2025-12-04T09:34:33.0819157Z inflating: build/bin/test_vec_half_AVX512 2025-12-04T09:34:33.0869116Z inflating: build/bin/test_vec_half_AVX2 2025-12-04T09:34:33.0918865Z inflating: build/bin/BackoffTest 2025-12-04T09:34:33.0970290Z inflating: build/bin/FileStoreTest 2025-12-04T09:34:33.1023109Z inflating: build/bin/static_runtime_bench 2025-12-04T09:34:33.1252061Z inflating: build/bin/static_runtime_test 2025-12-04T09:34:33.1319708Z inflating: build/bin/Dict_test 2025-12-04T09:34:33.1368661Z inflating: build/bin/Dimname_test 2025-12-04T09:34:33.1431597Z inflating: build/bin/MaybeOwned_test 2025-12-04T09:34:33.1486557Z inflating: build/bin/NamedTensor_test 2025-12-04T09:34:33.1545227Z inflating: build/bin/apply_utils_test 2025-12-04T09:34:33.1599965Z inflating: build/bin/atest 2025-12-04T09:34:33.1660674Z inflating: build/bin/basic 2025-12-04T09:34:33.1713627Z inflating: build/bin/broadcast_test 2025-12-04T09:34:33.1762905Z inflating: build/bin/cpu_allocator_test 2025-12-04T09:34:33.1818616Z inflating: build/bin/cpu_generator_test 2025-12-04T09:34:33.1869419Z inflating: build/bin/cpu_profiling_allocator_test 2025-12-04T09:34:33.1955327Z inflating: build/bin/cpu_rng_test 2025-12-04T09:34:33.2003829Z inflating: build/bin/dlconvertor_test 2025-12-04T09:34:33.2059620Z inflating: build/bin/extension_backend_test 2025-12-04T09:34:33.2114191Z inflating: build/bin/half_test 2025-12-04T09:34:33.2204349Z inflating: build/bin/ivalue_test 2025-12-04T09:34:33.2253240Z inflating: build/bin/lazy_tensor_test 2025-12-04T09:34:33.2303695Z inflating: build/bin/math_kernel_test 2025-12-04T09:34:33.2356334Z inflating: build/bin/memory_format_test 2025-12-04T09:34:33.2407564Z inflating: build/bin/memory_overlapping_test 2025-12-04T09:34:33.2459506Z inflating: build/bin/mobile_memory_cleanup 2025-12-04T09:34:33.2511740Z inflating: build/bin/native_test 2025-12-04T09:34:33.2561562Z inflating: build/bin/operator_name_test 2025-12-04T09:34:33.2611225Z inflating: build/bin/operators_test 2025-12-04T09:34:33.2661136Z inflating: build/bin/packedtensoraccessor_test 2025-12-04T09:34:33.2725652Z inflating: build/bin/pow_test 2025-12-04T09:34:33.2780464Z inflating: build/bin/quantized_test 2025-12-04T09:34:33.2827925Z inflating: build/bin/reduce_ops_test 2025-12-04T09:34:33.2876419Z inflating: build/bin/reportMemoryUsage_test 2025-12-04T09:34:33.2930473Z inflating: build/bin/scalar_tensor_test 2025-12-04T09:34:33.2987186Z inflating: build/bin/scalar_test 2025-12-04T09:34:33.3037621Z inflating: build/bin/StorageUtils_test 2025-12-04T09:34:33.3088378Z inflating: build/bin/stride_properties_test 2025-12-04T09:34:33.3160485Z inflating: build/bin/tensor_iterator_test 2025-12-04T09:34:33.3213188Z inflating: build/bin/test_parallel 2025-12-04T09:34:33.3263663Z inflating: build/bin/thread_init_test 2025-12-04T09:34:33.3315658Z inflating: build/bin/type_ptr_test 2025-12-04T09:34:33.3372826Z inflating: build/bin/type_test 2025-12-04T09:34:33.3422781Z inflating: build/bin/undefined_tensor_test 2025-12-04T09:34:33.3471006Z inflating: build/bin/verify_api_visibility 2025-12-04T09:34:33.3538145Z inflating: build/bin/legacy_vmap_test 2025-12-04T09:34:33.3587189Z inflating: build/bin/weakref_test 2025-12-04T09:34:33.3636337Z inflating: build/bin/wrapdim_test 2025-12-04T09:34:33.3687054Z inflating: build/bin/xla_tensor_test 2025-12-04T09:34:33.3745152Z inflating: build/bin/IListRef_test 2025-12-04T09:34:33.3841520Z inflating: build/bin/List_test 2025-12-04T09:34:33.3904899Z inflating: build/bin/KernelFunction_test 2025-12-04T09:34:33.4017697Z inflating: build/bin/kernel_function_legacy_test 2025-12-04T09:34:33.4102979Z inflating: build/bin/kernel_function_test 2025-12-04T09:34:33.4222550Z inflating: build/bin/kernel_lambda_legacy_test 2025-12-04T09:34:33.4313861Z inflating: build/bin/kernel_lambda_test 2025-12-04T09:34:33.4372434Z inflating: build/bin/kernel_stackbased_test 2025-12-04T09:34:33.4461723Z inflating: build/bin/make_boxed_from_unboxed_functor_test 2025-12-04T09:34:33.4510563Z inflating: build/bin/CppSignature_test 2025-12-04T09:34:33.4562623Z inflating: build/bin/backend_fallback_test 2025-12-04T09:34:33.4611413Z inflating: build/bin/op_allowlist_test 2025-12-04T09:34:33.4886898Z inflating: build/bin/op_registration_test 2025-12-04T09:34:33.4954091Z inflating: build/bin/inline_container_test 2025-12-04T09:34:33.5263616Z inflating: build/bin/test_lazy 2025-12-04T09:34:33.5317249Z inflating: build/bin/TCPStoreTest 2025-12-04T09:34:33.5368697Z inflating: build/bin/HashStoreTest 2025-12-04T09:34:33.6349334Z inflating: build/bin/test_jit 2025-12-04T09:34:33.6411437Z inflating: build/bin/ProcessGroupGlooTest 2025-12-04T09:34:33.6467436Z inflating: build/bin/test_aoti_inference 2025-12-04T09:34:33.6473239Z inflating: build/bin/example_allreduce 2025-12-04T09:34:33.6522872Z inflating: build/bin/test_dist_autograd 2025-12-04T09:34:33.6588134Z inflating: build/bin/test_cpp_rpc 2025-12-04T09:34:33.6593143Z inflating: build/bin/parallel_benchmark 2025-12-04T09:34:33.7617002Z inflating: build/bin/test_api 2025-12-04T09:34:33.7620065Z inflating: build/bin/torch_shm_manager 2025-12-04T09:34:33.7620658Z creating: .additional_ci_files/ 2025-12-04T09:34:33.7675378Z inflating: .additional_ci_files/test-times.json 2025-12-04T09:34:33.7878807Z inflating: .additional_ci_files/test-class-times.json 2025-12-04T09:34:33.7909361Z ##[group]Run rm artifacts.zip 2025-12-04T09:34:33.7909565Z rm artifacts.zip 2025-12-04T09:34:33.7914524Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:33.7914763Z env: 2025-12-04T09:34:33.7914906Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:33.7915077Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:33.7915241Z ##[endgroup] 2025-12-04T09:34:33.9138823Z ##[group]Run df -H 2025-12-04T09:34:33.9139002Z df -H 2025-12-04T09:34:33.9143420Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:33.9143672Z env: 2025-12-04T09:34:33.9143817Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:33.9143990Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:33.9144171Z ##[endgroup] 2025-12-04T09:34:33.9183244Z Filesystem Size Used Avail Use% Mounted on 2025-12-04T09:34:33.9183534Z devtmpfs 4.2M 0 4.2M 0% /dev 2025-12-04T09:34:33.9183759Z tmpfs 67G 0 67G 0% /dev/shm 2025-12-04T09:34:33.9183976Z tmpfs 27G 791k 27G 1% /run 2025-12-04T09:34:33.9184178Z /dev/nvme0n1p1 215G 72G 144G 34% / 2025-12-04T09:34:33.9184390Z tmpfs 67G 13k 67G 1% /tmp 2025-12-04T09:34:33.9184611Z /dev/nvme0n1p128 11M 1.4M 9.2M 13% /boot/efi 2025-12-04T09:34:33.9214002Z Prepare all required actions 2025-12-04T09:34:33.9214789Z Getting action download info 2025-12-04T09:34:34.0609749Z ##[group]Run ./.github/actions/download-td-artifacts 2025-12-04T09:34:34.0609985Z with: 2025-12-04T09:34:34.0610131Z env: 2025-12-04T09:34:34.0610273Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:34.0610452Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:34.0610617Z ##[endgroup] 2025-12-04T09:34:34.0949581Z ##[group]Run seemethere/download-artifact-s3@v4 2025-12-04T09:34:34.0949933Z with: 2025-12-04T09:34:34.0950078Z name: td_results 2025-12-04T09:34:34.0950247Z s3-bucket: gha-artifacts 2025-12-04T09:34:34.0950425Z region: us-east-1 2025-12-04T09:34:34.0950569Z env: 2025-12-04T09:34:34.0950714Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:34.0950886Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:34.0951049Z ##[endgroup] 2025-12-04T09:34:34.4650103Z (node:52440) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-12-04T09:34:34.4654003Z 2025-12-04T09:34:34.4654277Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-12-04T09:34:34.4654636Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-12-04T09:34:34.4654973Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-12-04T09:34:34.5446275Z Found 0 objects with prefix pytorch/pytorch/19923066595/td_results/ 2025-12-04T09:34:34.5452572Z Artifact download has finished successfully 2025-12-04T09:34:34.5631086Z ##[group]Run mkdir -p .additional_ci_files 2025-12-04T09:34:34.5631335Z mkdir -p .additional_ci_files 2025-12-04T09:34:34.5631614Z mv td_results.json .additional_ci_files/td_results.json || true 2025-12-04T09:34:34.5636004Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:34.5636296Z env: 2025-12-04T09:34:34.5636463Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:34.5636643Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:34.5636809Z ##[endgroup] 2025-12-04T09:34:34.5684463Z mv: cannot stat 'td_results.json': No such file or directory 2025-12-04T09:34:34.5723286Z ##[group]Run .github/scripts/parse_ref.py 2025-12-04T09:34:34.5723579Z .github/scripts/parse_ref.py 2025-12-04T09:34:34.5727397Z shell: /usr/bin/bash -e {0} 2025-12-04T09:34:34.5727587Z env: 2025-12-04T09:34:34.5727740Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:34.5727924Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:34.5728092Z ##[endgroup] 2025-12-04T09:34:34.5927327Z Setting output branch=main 2025-12-04T09:34:34.6020620Z Prepare all required actions 2025-12-04T09:34:34.6020939Z Getting action download info 2025-12-04T09:34:34.7325515Z ##[group]Run ./.github/actions/filter-test-configs 2025-12-04T09:34:34.7325753Z with: 2025-12-04T09:34:34.7326137Z github-token: *** 2025-12-04T09:34:34.7331907Z test-matrix: {"include": [{"config": "cpu_inductor_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_avx2_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_timm", "shard": 1, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_timm", "shard": 2, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}]} 2025-12-04T09:34:34.7337793Z job-name: periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx) 2025-12-04T09:34:34.7338210Z env: 2025-12-04T09:34:34.7338358Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:34.7338643Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:34.7338812Z ##[endgroup] 2025-12-04T09:34:34.7361799Z ##[group]Run nick-fields/retry@v3.0.0 2025-12-04T09:34:34.7362136Z with: 2025-12-04T09:34:34.7362298Z shell: bash 2025-12-04T09:34:34.7362463Z timeout_minutes: 10 2025-12-04T09:34:34.7362637Z max_attempts: 5 2025-12-04T09:34:34.7362802Z retry_wait_seconds: 30 2025-12-04T09:34:34.7363270Z command: set -eux # PyYAML 6.0 doesn't work with MacOS x86 anymore # This must run on Python-3.7 (AmazonLinux2) so can't use request=3.32.2 python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-12-04T09:34:34.7363726Z polling_interval_seconds: 1 2025-12-04T09:34:34.7363900Z warning_on_retry: true 2025-12-04T09:34:34.7364066Z continue_on_error: false 2025-12-04T09:34:34.7364228Z env: 2025-12-04T09:34:34.7364359Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:34.7364525Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:34.7364822Z GITHUB_TOKEN: *** 2025-12-04T09:34:34.7364995Z ##[endgroup] 2025-12-04T09:34:34.8203635Z + python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-12-04T09:34:34.9844064Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T09:34:35.0631279Z Collecting requests==2.27.1 2025-12-04T09:34:35.0766348Z Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB) 2025-12-04T09:34:35.1959402Z Collecting pyyaml==6.0.2 2025-12-04T09:34:35.2011285Z Downloading PyYAML-6.0.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (737 kB) 2025-12-04T09:34:35.2223349Z Requirement already satisfied: urllib3<1.27,>=1.21.1 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (1.25.10) 2025-12-04T09:34:35.2573415Z Collecting certifi>=2017.4.17 2025-12-04T09:34:35.2608816Z Downloading certifi-2025.11.12-py3-none-any.whl (159 kB) 2025-12-04T09:34:35.5170999Z Collecting charset-normalizer~=2.0.0 2025-12-04T09:34:35.5206062Z Downloading charset_normalizer-2.0.12-py3-none-any.whl (39 kB) 2025-12-04T09:34:35.6064742Z Requirement already satisfied: idna<4,>=2.5 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (2.10) 2025-12-04T09:34:35.6618144Z Installing collected packages: charset-normalizer, certifi, requests, pyyaml 2025-12-04T09:34:35.9185644Z Successfully installed certifi-2025.11.12 charset-normalizer-2.0.12 pyyaml-6.0.2 requests-2.27.1 2025-12-04T09:34:36.7937902Z Command completed after 1 attempt(s). 2025-12-04T09:34:36.7994131Z ##[group]Run set -x 2025-12-04T09:34:36.7994310Z set -x 2025-12-04T09:34:36.7994463Z  2025-12-04T09:34:36.7994703Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-12-04T09:34:36.7994986Z # in runner workspace 2025-12-04T09:34:36.7995232Z python3 "${GITHUB_ACTION_PATH}/../../scripts/parse_ref.py" 2025-12-04T09:34:36.7999810Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:36.8000040Z env: 2025-12-04T09:34:36.8000184Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:36.8000379Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:36.8000564Z ##[endgroup] 2025-12-04T09:34:36.8025487Z + python3 /home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/filter-test-configs/../../scripts/parse_ref.py 2025-12-04T09:34:36.8157169Z Setting output branch=main 2025-12-04T09:34:36.8209025Z ##[group]Run echo "Workflow: ${GITHUB_WORKFLOW}" 2025-12-04T09:34:36.8209308Z echo "Workflow: ${GITHUB_WORKFLOW}" 2025-12-04T09:34:36.8209516Z echo "Job name: ${JOB_NAME}" 2025-12-04T09:34:36.8209704Z  2025-12-04T09:34:36.8209937Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-12-04T09:34:36.8210229Z # in runner workspace 2025-12-04T09:34:36.8210489Z python3 "${GITHUB_ACTION_PATH}/../../scripts/filter_test_configs.py" \ 2025-12-04T09:34:36.8210771Z  --workflow "${GITHUB_WORKFLOW}" \ 2025-12-04T09:34:36.8210975Z  --job-name "${JOB_NAME}" \ 2025-12-04T09:34:36.8216737Z  --test-matrix "{"include": [{"config": "cpu_inductor_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_avx2_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_timm", "shard": 1, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_timm", "shard": 2, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}]}" \ 2025-12-04T09:34:36.8222759Z  --selected-test-configs "" \ 2025-12-04T09:34:36.8222993Z  --pr-number "${PR_NUMBER}" \ 2025-12-04T09:34:36.8223211Z  --tag "${TAG}" \ 2025-12-04T09:34:36.8223412Z  --event-name "${EVENT_NAME}" \ 2025-12-04T09:34:36.8223636Z  --schedule "${SCHEDULE}" \ 2025-12-04T09:34:36.8223855Z  --branch "${HEAD_BRANCH}" 2025-12-04T09:34:36.8228043Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:36.8228301Z env: 2025-12-04T09:34:36.8228530Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:36.8228726Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:36.8229304Z GITHUB_TOKEN: *** 2025-12-04T09:34:36.8229698Z JOB_NAME: periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx) 2025-12-04T09:34:36.8230098Z PR_NUMBER: 2025-12-04T09:34:36.8230439Z TAG: 2025-12-04T09:34:36.8230598Z EVENT_NAME: schedule 2025-12-04T09:34:36.8230787Z SCHEDULE: 45 0,4,8,12,16,20 * * 1-5 2025-12-04T09:34:36.8230984Z HEAD_BRANCH: main 2025-12-04T09:34:36.8231160Z ##[endgroup] 2025-12-04T09:34:36.8255908Z Workflow: inductor-periodic 2025-12-04T09:34:36.8256363Z Job name: periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx) 2025-12-04T09:34:37.0529123Z Setting output keep-going=True 2025-12-04T09:34:37.0529635Z Setting output ci-verbose-test-logs=False 2025-12-04T09:34:37.0530547Z Setting output ci-test-showlocals=False 2025-12-04T09:34:37.0530877Z Setting output ci-no-test-timeout=False 2025-12-04T09:34:37.0531138Z Setting output ci-no-td=False 2025-12-04T09:34:37.0531339Z Setting output ci-td-distributed=False 2025-12-04T09:34:37.0531553Z Setting output is-unstable=False 2025-12-04T09:34:37.0531746Z Setting output reenabled-issues= 2025-12-04T09:34:37.0537735Z Setting output test-matrix={"include": [{"config": "cpu_inductor_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_avx2_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_timm", "shard": 1, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_timm", "shard": 2, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}]} 2025-12-04T09:34:37.0543840Z Setting output is-test-matrix-empty=False 2025-12-04T09:34:37.0647644Z ##[group]Run echo "Filtered matrix:" 2025-12-04T09:34:37.0647901Z echo "Filtered matrix:" 2025-12-04T09:34:37.0653673Z echo "{"include": [{"config": "cpu_inductor_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_inductor_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_avx2_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_timm", "shard": 1, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_avx2_timm", "shard": 2, "num_shards": 2, "runner": "linux.10xlarge.avx2"}, {"config": "cpu_inductor_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_huggingface", "shard": 1, "num_shards": 1, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_timm", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_timm", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "cpu_aot_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_amp_freezing_torchbench", "shard": 1, "num_shards": 2, "runner": "linux.8xlarge.amx"}, {"config": "dynamic_cpu_aot_inductor_amp_freezing_torchbench", "shard": 2, "num_shards": 2, "runner": "linux.8xlarge.amx"}]}" 2025-12-04T09:34:37.0659587Z  2025-12-04T09:34:37.0659730Z echo 2025-12-04T09:34:37.0659915Z echo "Is the current job unstable? False" 2025-12-04T09:34:37.0660123Z  2025-12-04T09:34:37.0660271Z echo 2025-12-04T09:34:37.0660456Z echo "Is keep-going label set? True" 2025-12-04T09:34:37.0660663Z  2025-12-04T09:34:37.0660810Z echo 2025-12-04T09:34:37.0660978Z echo "Reenabled issues? " 2025-12-04T09:34:37.0665372Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:37.0665603Z env: 2025-12-04T09:34:37.0665753Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:37.0665927Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:37.0666082Z ##[endgroup] 2025-12-04T09:34:37.0687894Z Filtered matrix: 2025-12-04T09:34:37.0697764Z {include: [{config: cpu_inductor_huggingface, shard: 1, num_shards: 1, runner: linux.8xlarge.amx}, {config: cpu_inductor_timm, shard: 1, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_inductor_timm, shard: 2, num_shards: 2, runner: linux.8xlarge.amx}, {config: dynamic_cpu_inductor_huggingface, shard: 1, num_shards: 1, runner: linux.8xlarge.amx}, {config: dynamic_cpu_inductor_timm, shard: 1, num_shards: 2, runner: linux.8xlarge.amx}, {config: dynamic_cpu_inductor_timm, shard: 2, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_inductor_freezing_avx2_huggingface, shard: 1, num_shards: 1, runner: linux.10xlarge.avx2}, {config: cpu_inductor_freezing_avx2_torchbench, shard: 1, num_shards: 2, runner: linux.10xlarge.avx2}, {config: cpu_inductor_freezing_avx2_torchbench, shard: 2, num_shards: 2, runner: linux.10xlarge.avx2}, {config: cpu_inductor_freezing_avx2_timm, shard: 1, num_shards: 2, runner: linux.10xlarge.avx2}, {config: cpu_inductor_freezing_avx2_timm, shard: 2, num_shards: 2, runner: linux.10xlarge.avx2}, {config: cpu_inductor_freezing_huggingface, shard: 1, num_shards: 1, runner: linux.8xlarge.amx}, {config: cpu_inductor_freezing_timm, shard: 1, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_inductor_freezing_timm, shard: 2, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_inductor_freezing_torchbench, shard: 1, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_inductor_freezing_torchbench, shard: 2, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_inductor_amp_freezing_huggingface, shard: 1, num_shards: 1, runner: linux.8xlarge.amx}, {config: cpu_inductor_amp_freezing_timm, shard: 1, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_inductor_amp_freezing_timm, shard: 2, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_inductor_amp_freezing_torchbench, shard: 1, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_inductor_amp_freezing_torchbench, shard: 2, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_aot_inductor_freezing_huggingface, shard: 1, num_shards: 1, runner: linux.8xlarge.amx}, {config: cpu_aot_inductor_freezing_timm, shard: 1, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_aot_inductor_freezing_timm, shard: 2, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_aot_inductor_freezing_torchbench, shard: 1, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_aot_inductor_freezing_torchbench, shard: 2, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_aot_inductor_amp_freezing_torchbench, shard: 1, num_shards: 2, runner: linux.8xlarge.amx}, {config: cpu_aot_inductor_amp_freezing_torchbench, shard: 2, num_shards: 2, runner: linux.8xlarge.amx}, {config: dynamic_cpu_aot_inductor_freezing_torchbench, shard: 1, num_shards: 2, runner: linux.8xlarge.amx}, {config: dynamic_cpu_aot_inductor_freezing_torchbench, shard: 2, num_shards: 2, runner: linux.8xlarge.amx}, {config: dynamic_cpu_aot_inductor_amp_freezing_torchbench, shard: 1, num_shards: 2, runner: linux.8xlarge.amx}, {config: dynamic_cpu_aot_inductor_amp_freezing_torchbench, shard: 2, num_shards: 2, runner: linux.8xlarge.amx}]} 2025-12-04T09:34:37.0703588Z 2025-12-04T09:34:37.0703676Z Is the current job unstable? False 2025-12-04T09:34:37.0703820Z 2025-12-04T09:34:37.0703903Z Is keep-going label set? True 2025-12-04T09:34:37.0704032Z 2025-12-04T09:34:37.0704107Z Reenabled issues? 2025-12-04T09:34:37.0768059Z ##[group]Run echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-12-04T09:34:37.0768416Z echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-12-04T09:34:37.0772448Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:37.0772676Z env: 2025-12-04T09:34:37.0772832Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:37.0773013Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:37.0773185Z JOB_TIMEOUT: 240 2025-12-04T09:34:37.0773342Z ##[endgroup] 2025-12-04T09:34:37.0995253Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:34:37.0995578Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:34:37.0995855Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:34:37.0999553Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:34:37.0999800Z env: 2025-12-04T09:34:37.0999958Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:37.1000130Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:37.1000288Z ##[endgroup] 2025-12-04T09:34:37.1116488Z ##[group]Run set -x 2025-12-04T09:34:37.1116720Z set -x 2025-12-04T09:34:37.1116875Z  2025-12-04T09:34:37.1117052Z if [[ $TEST_CONFIG == 'multigpu' ]]; then 2025-12-04T09:34:37.1117295Z  TEST_COMMAND=.ci/pytorch/multigpu-test.sh 2025-12-04T09:34:37.1117547Z elif [[ $BUILD_ENVIRONMENT == *onnx* ]]; then 2025-12-04T09:34:37.1117874Z  TEST_COMMAND=.ci/onnx/test.sh 2025-12-04T09:34:37.1118059Z else 2025-12-04T09:34:37.1118230Z  TEST_COMMAND=.ci/pytorch/test.sh 2025-12-04T09:34:37.1118424Z fi 2025-12-04T09:34:37.1118557Z  2025-12-04T09:34:37.1118732Z # Leaving 1GB for the runner and other things 2025-12-04T09:34:37.1119081Z TOTAL_AVAILABLE_MEMORY_IN_GB=$(awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo) 2025-12-04T09:34:37.1119593Z # https://docs.docker.com/engine/containers/resource_constraints/#--memory-swap-details, the 3GB swap 2025-12-04T09:34:37.1120000Z # comes from https://github.com/pytorch/test-infra/pull/6058 2025-12-04T09:34:37.1120312Z TOTAL_MEMORY_WITH_SWAP=$(("${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}" + 3)) 2025-12-04T09:34:37.1120556Z  2025-12-04T09:34:37.1120730Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-12-04T09:34:37.1120948Z  SHM_OPTS= 2025-12-04T09:34:37.1121116Z  JENKINS_USER= 2025-12-04T09:34:37.1121335Z  # ensure that docker container cleanly exits in 12 hours 2025-12-04T09:34:37.1121620Z  # if for some reason cleanup action doesn't stop container 2025-12-04T09:34:37.1121965Z  # when job is cancelled 2025-12-04T09:34:37.1122190Z  DOCKER_SHELL_CMD="sleep 12h" 2025-12-04T09:34:37.1122411Z  USED_IMAGE="${DOCKER_IMAGE_S390X}" 2025-12-04T09:34:37.1122622Z else 2025-12-04T09:34:37.1122804Z  SHM_OPTS="--shm-size=${SHM_SIZE}" 2025-12-04T09:34:37.1123032Z  JENKINS_USER="--user jenkins" 2025-12-04T09:34:37.1123254Z  DOCKER_SHELL_CMD= 2025-12-04T09:34:37.1123446Z  USED_IMAGE="${DOCKER_IMAGE}" 2025-12-04T09:34:37.1123626Z fi 2025-12-04T09:34:37.1123771Z  2025-12-04T09:34:37.1123990Z # detached container should get cleaned up by teardown_ec2_linux 2025-12-04T09:34:37.1124308Z # TODO: Stop building test binaries as part of the build phase 2025-12-04T09:34:37.1124651Z # Used for GPU_FLAG, SHM_OPTS, JENKINS_USER and DOCKER_SHELL_CMD since that doesn't play nice 2025-12-04T09:34:37.1124962Z # shellcheck disable=SC2086,SC2090 2025-12-04T09:34:37.1125177Z container_name=$(docker run \ 2025-12-04T09:34:37.1125371Z  ${GPU_FLAG:-} \ 2025-12-04T09:34:37.1125569Z  ${SCCACHE_SERVER_PORT_DOCKER_FLAG:-} \ 2025-12-04T09:34:37.1125786Z  -e BUILD_ENVIRONMENT \ 2025-12-04T09:34:37.1125977Z  -e PR_NUMBER \ 2025-12-04T09:34:37.1126154Z  -e GITHUB_ACTIONS \ 2025-12-04T09:34:37.1126342Z  -e GITHUB_REPOSITORY \ 2025-12-04T09:34:37.1126532Z  -e GITHUB_WORKFLOW \ 2025-12-04T09:34:37.1126711Z  -e GITHUB_JOB \ 2025-12-04T09:34:37.1126891Z  -e GITHUB_RUN_ID \ 2025-12-04T09:34:37.1127072Z  -e GITHUB_RUN_NUMBER \ 2025-12-04T09:34:37.1127260Z  -e GITHUB_RUN_ATTEMPT \ 2025-12-04T09:34:37.1127450Z  -e JOB_ID \ 2025-12-04T09:34:37.1127620Z  -e JOB_NAME \ 2025-12-04T09:34:37.1127791Z  -e BASE_SHA \ 2025-12-04T09:34:37.1127953Z  -e BRANCH \ 2025-12-04T09:34:37.1128113Z  -e SHA1 \ 2025-12-04T09:34:37.1128280Z  -e AWS_DEFAULT_REGION \ 2025-12-04T09:34:37.1128462Z  -e IN_WHEEL_TEST \ 2025-12-04T09:34:37.1128639Z  -e SHARD_NUMBER \ 2025-12-04T09:34:37.1128816Z  -e TEST_CONFIG \ 2025-12-04T09:34:37.1128987Z  -e NUM_TEST_SHARDS \ 2025-12-04T09:34:37.1129255Z  -e REENABLED_ISSUES \ 2025-12-04T09:34:37.1129458Z  -e CONTINUE_THROUGH_ERROR \ 2025-12-04T09:34:37.1129650Z  -e VERBOSE_TEST_LOGS \ 2025-12-04T09:34:37.1129836Z  -e TEST_SHOWLOCALS \ 2025-12-04T09:34:37.1130022Z  -e NO_TEST_TIMEOUT \ 2025-12-04T09:34:37.1130445Z  -e NO_TD \ 2025-12-04T09:34:37.1130690Z  -e TD_DISTRIBUTED \ 2025-12-04T09:34:37.1130874Z  -e PR_LABELS \ 2025-12-04T09:34:37.1131078Z  -e MAX_JOBS="$(nproc --ignore=2)" \ 2025-12-04T09:34:37.1131294Z  -e SCCACHE_BUCKET \ 2025-12-04T09:34:37.1131483Z  -e SCCACHE_REGION \ 2025-12-04T09:34:37.1131666Z  -e XLA_CUDA \ 2025-12-04T09:34:37.1131853Z  -e XLA_CLANG_CACHE_S3_BUCKET_NAME \ 2025-12-04T09:34:37.1132107Z  -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK \ 2025-12-04T09:34:37.1132354Z  -e PYTORCH_TEST_RERUN_DISABLED_TESTS \ 2025-12-04T09:34:37.1132592Z  -e SKIP_SCCACHE_INITIALIZATION=1 \ 2025-12-04T09:34:37.1132806Z  -e HUGGING_FACE_HUB_TOKEN \ 2025-12-04T09:34:37.1133016Z  -e VLLM_TEST_HUGGING_FACE_TOKEN \ 2025-12-04T09:34:37.1133225Z  -e SCRIBE_GRAPHQL_ACCESS_TOKEN \ 2025-12-04T09:34:37.1133420Z  -e DASHBOARD_TAG \ 2025-12-04T09:34:37.1133596Z  -e ARTIFACTS_FILE_SUFFIX \ 2025-12-04T09:34:37.1133822Z  --memory="${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}g" \ 2025-12-04T09:34:37.1134071Z  --memory-swap="${TOTAL_MEMORY_WITH_SWAP}g" \ 2025-12-04T09:34:37.1134321Z  --env-file="/tmp/github_env_${GITHUB_RUN_ID}" \ 2025-12-04T09:34:37.1134554Z  --security-opt seccomp=unconfined \ 2025-12-04T09:34:37.1134762Z  --cap-add=SYS_PTRACE \ 2025-12-04T09:34:37.1134948Z  --ipc=host \ 2025-12-04T09:34:37.1135108Z  ${SHM_OPTS} \ 2025-12-04T09:34:37.1135271Z  --tty \ 2025-12-04T09:34:37.1135427Z  --detach \ 2025-12-04T09:34:37.1135597Z  --name="${container_name}" \ 2025-12-04T09:34:37.1135789Z  ${JENKINS_USER} \ 2025-12-04T09:34:37.1136005Z  -v "${GITHUB_WORKSPACE}:/var/lib/jenkins/workspace" \ 2025-12-04T09:34:37.1136241Z  -w /var/lib/jenkins/workspace \ 2025-12-04T09:34:37.1136428Z  "${USED_IMAGE}" \ 2025-12-04T09:34:37.1136600Z  ${DOCKER_SHELL_CMD} 2025-12-04T09:34:37.1136769Z ) 2025-12-04T09:34:37.1136970Z echo "DOCKER_CONTAINER_ID=${container_name}" >> "${GITHUB_ENV}" 2025-12-04T09:34:37.1137211Z  2025-12-04T09:34:37.1137383Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-12-04T09:34:37.1137721Z  docker exec -t "${container_name}" sh -c "python3 -m pip install -r .ci/docker/requirements-ci.txt" 2025-12-04T09:34:37.1138011Z fi 2025-12-04T09:34:37.1138148Z  2025-12-04T09:34:37.1138438Z docker exec -t "${container_name}" sh -c "python3 -m pip install $(echo dist/*.whl)[opt-einsum] && ${TEST_COMMAND}" 2025-12-04T09:34:37.1142207Z shell: /usr/bin/bash -e {0} 2025-12-04T09:34:37.1142390Z env: 2025-12-04T09:34:37.1142543Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:34:37.1142714Z HAS_NVIDIA_GPU: false 2025-12-04T09:34:37.1142918Z BUILD_ENVIRONMENT: linux-jammy-py3.10-gcc11-build 2025-12-04T09:34:37.1143140Z PR_NUMBER: 2025-12-04T09:34:37.1143302Z GITHUB_REPOSITORY: pytorch/pytorch 2025-12-04T09:34:37.1143508Z GITHUB_WORKFLOW: inductor-periodic 2025-12-04T09:34:37.1143699Z GITHUB_JOB: test 2025-12-04T09:34:37.1143856Z GITHUB_RUN_ID: 19923066595 2025-12-04T09:34:37.1144025Z GITHUB_RUN_NUMBER: 67027 2025-12-04T09:34:37.1144195Z GITHUB_RUN_ATTEMPT: 1 2025-12-04T09:34:37.1144356Z JOB_ID: 57118563344 2025-12-04T09:34:37.1144691Z JOB_NAME: periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx) 2025-12-04T09:34:37.1145051Z BRANCH: main 2025-12-04T09:34:37.1145322Z SHA1: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:34:37.1145561Z BASE_SHA: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:34:37.1145801Z TEST_CONFIG: cpu_inductor_amp_freezing_huggingface 2025-12-04T09:34:37.1146013Z SHARD_NUMBER: 1 2025-12-04T09:34:37.1146167Z NUM_TEST_SHARDS: 1 2025-12-04T09:34:37.1146313Z EXTRA_FLAGS: 2025-12-04T09:34:37.1146472Z OP_BENCHMARK_TESTS: 2025-12-04T09:34:37.1146680Z REENABLED_ISSUES: 2025-12-04T09:34:37.1146834Z CONTINUE_THROUGH_ERROR: True 2025-12-04T09:34:37.1147010Z VERBOSE_TEST_LOGS: False 2025-12-04T09:34:37.1147181Z TEST_SHOWLOCALS: False 2025-12-04T09:34:37.1147342Z NO_TEST_TIMEOUT: False 2025-12-04T09:34:37.1147503Z NO_TD: False 2025-12-04T09:34:37.1147652Z TD_DISTRIBUTED: False 2025-12-04T09:34:37.1147844Z SCCACHE_BUCKET: ossci-compiler-cache-circleci-v2 2025-12-04T09:34:37.1148067Z SCCACHE_REGION: us-east-1 2025-12-04T09:34:37.1148236Z SHM_SIZE: 1g 2025-12-04T09:34:37.1148703Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:34:37.1149495Z DOCKER_IMAGE_S390X: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:34:37.1149983Z XLA_CUDA: 2025-12-04T09:34:37.1150207Z XLA_CLANG_CACHE_S3_BUCKET_NAME: ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:34:37.1150482Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: 0 2025-12-04T09:34:37.1150677Z PYTORCH_TEST_RERUN_DISABLED_TESTS: 0 2025-12-04T09:34:37.1150860Z DASHBOARD_TAG: 2025-12-04T09:34:37.1151178Z VLLM_TEST_HUGGING_FACE_TOKEN: *** 2025-12-04T09:34:37.1151433Z HUGGING_FACE_HUB_TOKEN: *** 2025-12-04T09:34:37.1151690Z SCRIBE_GRAPHQL_ACCESS_TOKEN: *** 2025-12-04T09:34:37.1152013Z ARTIFACTS_FILE_SUFFIX: test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344 2025-12-04T09:34:37.1152329Z ##[endgroup] 2025-12-04T09:34:37.1173268Z + [[ cpu_inductor_amp_freezing_huggingface == \m\u\l\t\i\g\p\u ]] 2025-12-04T09:34:37.1173617Z + [[ linux-jammy-py3.10-gcc11-build == *onnx* ]] 2025-12-04T09:34:37.1177864Z + TEST_COMMAND=.ci/pytorch/test.sh 2025-12-04T09:34:37.1178187Z ++ awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo 2025-12-04T09:34:37.1193819Z + TOTAL_AVAILABLE_MEMORY_IN_GB='122.780 ' 2025-12-04T09:34:37.1194251Z + TOTAL_MEMORY_WITH_SWAP=125 2025-12-04T09:34:37.1194616Z + [[ linux-jammy-py3.10-gcc11-build == *\s\3\9\0\x* ]] 2025-12-04T09:34:37.1194973Z + SHM_OPTS=--shm-size=1g 2025-12-04T09:34:37.1195259Z + JENKINS_USER='--user jenkins' 2025-12-04T09:34:37.1195547Z + DOCKER_SHELL_CMD= 2025-12-04T09:34:37.1196475Z + USED_IMAGE=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:34:37.1204972Z +++ nproc --ignore=2 2025-12-04T09:34:37.1381675Z ++ docker run -e BUILD_ENVIRONMENT -e PR_NUMBER -e GITHUB_ACTIONS -e GITHUB_REPOSITORY -e GITHUB_WORKFLOW -e GITHUB_JOB -e GITHUB_RUN_ID -e GITHUB_RUN_NUMBER -e GITHUB_RUN_ATTEMPT -e JOB_ID -e JOB_NAME -e BASE_SHA -e BRANCH -e SHA1 -e AWS_DEFAULT_REGION -e IN_WHEEL_TEST -e SHARD_NUMBER -e TEST_CONFIG -e NUM_TEST_SHARDS -e REENABLED_ISSUES -e CONTINUE_THROUGH_ERROR -e VERBOSE_TEST_LOGS -e TEST_SHOWLOCALS -e NO_TEST_TIMEOUT -e NO_TD -e TD_DISTRIBUTED -e PR_LABELS -e MAX_JOBS=30 -e SCCACHE_BUCKET -e SCCACHE_REGION -e XLA_CUDA -e XLA_CLANG_CACHE_S3_BUCKET_NAME -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK -e PYTORCH_TEST_RERUN_DISABLED_TESTS -e SKIP_SCCACHE_INITIALIZATION=1 -e HUGGING_FACE_HUB_TOKEN -e VLLM_TEST_HUGGING_FACE_TOKEN -e SCRIBE_GRAPHQL_ACCESS_TOKEN -e DASHBOARD_TAG -e ARTIFACTS_FILE_SUFFIX --memory=122g --memory-swap=125g --env-file=/tmp/github_env_19923066595 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --ipc=host --shm-size=1g --tty --detach --name= --user jenkins -v /home/ec2-user/actions-runner/_work/pytorch/pytorch:/var/lib/jenkins/workspace -w /var/lib/jenkins/workspace 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:34:50.9788068Z + container_name=25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:34:50.9788636Z + echo DOCKER_CONTAINER_ID=25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:34:50.9789286Z + [[ linux-jammy-py3.10-gcc11-build == *\s\3\9\0\x* ]] 2025-12-04T09:34:50.9794614Z ++ echo dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl 2025-12-04T09:34:50.9795364Z + docker exec -t 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 sh -c 'python3 -m pip install dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl[opt-einsum] && .ci/pytorch/test.sh' 2025-12-04T09:34:51.3539040Z Processing ./dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl (from torch==2.10.0a0+gitffd9b0f) 2025-12-04T09:34:51.5554604Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.18.0) 2025-12-04T09:34:51.5557500Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (4.15.0) 2025-12-04T09:34:51.5558246Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.13.3) 2025-12-04T09:34:51.5562776Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (2.8.8) 2025-12-04T09:34:51.5563613Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.1.6) 2025-12-04T09:34:51.5564475Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (2025.10.0) 2025-12-04T09:34:51.5575615Z Requirement already satisfied: opt-einsum>=3.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.3.0) 2025-12-04T09:34:51.5836453Z Requirement already satisfied: numpy>=1.7 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from opt-einsum>=3.3->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.22.4) 2025-12-04T09:34:51.5851043Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from sympy>=1.13.3->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.3.0) 2025-12-04T09:34:51.5895222Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from jinja2->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.0.3) 2025-12-04T09:34:52.3122411Z Installing collected packages: torch 2025-12-04T09:34:59.3187323Z ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts. 2025-12-04T09:34:59.3191833Z dall-e 0.1 requires torchvision, which is not installed. 2025-12-04T09:34:59.3196576Z effdet 0.4.1 requires torchvision, which is not installed. 2025-12-04T09:34:59.3196964Z python-doctr 1.0.0 requires torchvision>=0.15.0, which is not installed. 2025-12-04T09:34:59.3197337Z pytorch-labs-segment-anything-fast 0.2 requires torchao, which is not installed. 2025-12-04T09:34:59.3197784Z pytorch-labs-segment-anything-fast 0.2 requires torchvision>=0.17.0.dev20231026, which is not installed. 2025-12-04T09:34:59.3201242Z timm 1.0.22 requires torchvision, which is not installed. 2025-12-04T09:34:59.3201867Z Successfully installed torch-2.10.0a0+gitffd9b0f 2025-12-04T09:34:59.4177876Z + export TERM=vt100 2025-12-04T09:34:59.4179472Z + TERM=vt100 2025-12-04T09:34:59.4186206Z ++ dirname .ci/pytorch/test.sh 2025-12-04T09:34:59.4190553Z + source .ci/pytorch/common.sh 2025-12-04T09:34:59.4190858Z +++ dirname .ci/pytorch/common.sh 2025-12-04T09:34:59.4197381Z ++ source .ci/pytorch/common_utils.sh 2025-12-04T09:34:59.4197650Z +++ declare -f -t trap_add 2025-12-04T09:34:59.4201769Z ++ set -ex -o pipefail 2025-12-04T09:34:59.4202192Z ++ [[ linux-jammy-py3.10-gcc11-build == *rocm* ]] 2025-12-04T09:34:59.4202707Z ++ BUILD_TEST_LIBTORCH=0 2025-12-04T09:34:59.4219865Z ++ dirname .ci/pytorch/test.sh 2025-12-04T09:34:59.4238872Z + source .ci/pytorch/common-build.sh 2025-12-04T09:34:59.4242169Z ++ [[ linux-jammy-py3.10-gcc11-build != *win-* ]] 2025-12-04T09:34:59.4252704Z ++++ dirname .ci/pytorch/common-build.sh 2025-12-04T09:34:59.4257198Z +++ cd .ci/pytorch 2025-12-04T09:34:59.4262098Z +++ pwd -P 2025-12-04T09:34:59.4265013Z ++ script_dir=/var/lib/jenkins/workspace/.ci/pytorch 2025-12-04T09:34:59.4265390Z ++ [[ linux-jammy-py3.10-gcc11-build == *-pch* ]] 2025-12-04T09:34:59.4269606Z ++ which sccache 2025-12-04T09:34:59.4288010Z ++ [[ -z ossci-compiler-cache-circleci-v2 ]] 2025-12-04T09:34:59.4288356Z ++ sccache --stop-server 2025-12-04T09:34:59.4312107Z ++ true 2025-12-04T09:34:59.4314155Z ++ rm -f /var/lib/jenkins/sccache_error.log 2025-12-04T09:34:59.4323489Z ++ trap_add sccache_epilogue EXIT 2025-12-04T09:34:59.4323713Z ++ trap_add_cmd=sccache_epilogue 2025-12-04T09:34:59.4323920Z ++ shift 2025-12-04T09:34:59.4324085Z ++ for trap_add_name in "$@" 2025-12-04T09:34:59.4329159Z ++++ trap -p EXIT 2025-12-04T09:34:59.4339748Z +++ eval 'extract_trap_cmd ' 2025-12-04T09:34:59.4341779Z ++++ extract_trap_cmd 2025-12-04T09:34:59.4342101Z ++++ printf '%s\n' '' 2025-12-04T09:34:59.4347321Z +++ printf '%s\n' sccache_epilogue 2025-12-04T09:34:59.4349229Z ++ trap -- ' 2025-12-04T09:34:59.4349520Z sccache_epilogue' EXIT 2025-12-04T09:34:59.4353294Z ++ [[ -n 1 ]] 2025-12-04T09:34:59.4353691Z ++ echo 'Skipping sccache server initialization, setting environment variables' 2025-12-04T09:34:59.4359098Z Skipping sccache server initialization, setting environment variables 2025-12-04T09:34:59.4359426Z ++ export SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:34:59.4359657Z ++ SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:34:59.4359892Z ++ export SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:34:59.4360175Z ++ SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:34:59.4370988Z ++ export RUST_LOG=sccache::server=error 2025-12-04T09:34:59.4371263Z ++ RUST_LOG=sccache::server=error 2025-12-04T09:34:59.4371476Z ++ sccache --zero-stats 2025-12-04T09:34:59.6138549Z Statistics zeroed. 2025-12-04T09:34:59.6144507Z ++ which ccache 2025-12-04T09:34:59.6166153Z + [[ linux-jammy-py3.10-gcc11-build != *rocm* ]] 2025-12-04T09:34:59.6166565Z + [[ linux-jammy-py3.10-gcc11-build != *s390x* ]] 2025-12-04T09:34:59.6171332Z + [[ -d /var/lib/jenkins/workspace ]] 2025-12-04T09:34:59.6171754Z ++ stat -c %u /var/lib/jenkins/workspace 2025-12-04T09:34:59.6185643Z + WORKSPACE_ORIGINAL_OWNER_ID=1000 2025-12-04T09:34:59.6190227Z + trap_add cleanup_workspace EXIT 2025-12-04T09:34:59.6194859Z + trap_add_cmd=cleanup_workspace 2025-12-04T09:34:59.6196802Z + shift 2025-12-04T09:34:59.6197138Z + for trap_add_name in "$@" 2025-12-04T09:34:59.6197426Z +++ trap -p EXIT 2025-12-04T09:34:59.6197714Z ++ eval 'extract_trap_cmd trap -- '\'' 2025-12-04T09:34:59.6198010Z sccache_epilogue'\'' EXIT' 2025-12-04T09:34:59.6198232Z +++ extract_trap_cmd trap -- ' 2025-12-04T09:34:59.6198416Z sccache_epilogue' EXIT 2025-12-04T09:34:59.6198587Z +++ printf '%s\n' ' 2025-12-04T09:34:59.6198757Z sccache_epilogue' 2025-12-04T09:34:59.6198923Z ++ printf '%s\n' cleanup_workspace 2025-12-04T09:34:59.6199116Z + trap -- ' 2025-12-04T09:34:59.6199266Z sccache_epilogue 2025-12-04T09:34:59.6199420Z cleanup_workspace' EXIT 2025-12-04T09:34:59.6199627Z + sudo chown -R jenkins /var/lib/jenkins/workspace 2025-12-04T09:35:00.0410416Z + git config --global --add safe.directory /var/lib/jenkins/workspace 2025-12-04T09:35:00.0429326Z + [[ linux-jammy-py3.10-gcc11-build == *cuda* ]] 2025-12-04T09:35:00.0429897Z + echo 'Environment variables:' 2025-12-04T09:35:00.0430102Z Environment variables: 2025-12-04T09:35:00.0430472Z + env 2025-12-04T09:35:00.0439100Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:35:00.0439468Z CONTINUE_THROUGH_ERROR=True 2025-12-04T09:35:00.0439689Z BUILD_ENVIRONMENT=linux-jammy-py3.10-gcc11-build 2025-12-04T09:35:00.0440385Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-12-04T09:35:00.0440582Z HOSTNAME=25f649d00d6e 2025-12-04T09:35:00.0440948Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_53a1234b-4002-441a-a1d2-f0fc0a79a49e 2025-12-04T09:35:00.0441326Z GITHUB_ACTION=__run_3 2025-12-04T09:35:00.0441518Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-12-04T09:35:00.0441727Z GITHUB_RUN_NUMBER=67027 2025-12-04T09:35:00.0442080Z TEST_CONFIG=cpu_inductor_amp_freezing_huggingface 2025-12-04T09:35:00.0442321Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-12-04T09:35:00.0442560Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-12-04T09:35:00.0442799Z SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:35:00.0443126Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-12-04T09:35:00.0443353Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-12-04T09:35:00.0443588Z GITHUB_REF_TYPE=branch 2025-12-04T09:35:00.0443802Z BASE_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:35:00.0444015Z XLA_CUDA= 2025-12-04T09:35:00.0444180Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-12-04T09:35:00.0444434Z HUGGING_FACE_HUB_TOKEN=*** 2025-12-04T09:35:00.0444862Z *** 2025-12-04T09:35:00.0445037Z GITHUB_REPOSITORY_ID=65600975 2025-12-04T09:35:00.0445248Z GITHUB_ACTIONS=true 2025-12-04T09:35:00.0445464Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:35:00.0445740Z SHA1=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:35:00.0446014Z GITHUB_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:35:00.0446392Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/inductor-periodic.yml@refs/heads/main 2025-12-04T09:35:00.0446754Z UCC_HOME=/usr 2025-12-04T09:35:00.0446936Z VERBOSE_TEST_LOGS=False 2025-12-04T09:35:00.0447120Z GITHUB_REF=refs/heads/main 2025-12-04T09:35:00.0447311Z SHARD_NUMBER=1 2025-12-04T09:35:00.0447488Z GITHUB_REF_PROTECTED=true 2025-12-04T09:35:00.0447674Z HOME=/var/lib/jenkins 2025-12-04T09:35:00.0447883Z GITHUB_API_URL=https://api.github.com 2025-12-04T09:35:00.0448118Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-12-04T09:35:00.0448327Z UCX_COMMIT= 2025-12-04T09:35:00.0448480Z USE_SYSTEM_NCCL=1 2025-12-04T09:35:00.0448650Z NUM_TEST_SHARDS=1 2025-12-04T09:35:00.0448818Z UCX_HOME=/usr 2025-12-04T09:35:00.0449185Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_53a1234b-4002-441a-a1d2-f0fc0a79a49e 2025-12-04T09:35:00.0449799Z JOB_NAME=periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx) 2025-12-04T09:35:00.0450405Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_53a1234b-4002-441a-a1d2-f0fc0a79a49e 2025-12-04T09:35:00.0450910Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-12-04T09:35:00.0451245Z GITHUB_EVENT_NAME=schedule 2025-12-04T09:35:00.0451435Z DASHBOARD_TAG= 2025-12-04T09:35:00.0451606Z GITHUB_RUN_ID=19923066595 2025-12-04T09:35:00.0451789Z INSTALLED_OPENBLAS= 2025-12-04T09:35:00.0452184Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_53a1234b-4002-441a-a1d2-f0fc0a79a49e 2025-12-04T09:35:00.0452596Z GITHUB_ACTOR=pytorchmergebot 2025-12-04T09:35:00.0452770Z PR_NUMBER= 2025-12-04T09:35:00.0452919Z DESIRED_CUDA= 2025-12-04T09:35:00.0453075Z GITHUB_RUN_ATTEMPT=1 2025-12-04T09:35:00.0453242Z ANACONDA_PYTHON_VERSION=3.10 2025-12-04T09:35:00.0453461Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-12-04T09:35:00.0453680Z TERM=vt100 2025-12-04T09:35:00.0453823Z INSTALLED_VISION=yes 2025-12-04T09:35:00.0453988Z BRANCH=main 2025-12-04T09:35:00.0454142Z SCCACHE_REGION=us-east-1 2025-12-04T09:35:00.0454420Z OPENSSL_ROOT_DIR=/opt/openssl 2025-12-04T09:35:00.0454618Z BUILD_AOT_INDUCTOR_TEST=yes 2025-12-04T09:35:00.0454807Z CUDA_PATH=/usr/local/cuda 2025-12-04T09:35:00.0455127Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-12-04T09:35:00.0455469Z GITHUB_SERVER_URL=https://github.com 2025-12-04T09:35:00.0455671Z UCC_COMMIT= 2025-12-04T09:35:00.0455877Z REENABLED_ISSUES= 2025-12-04T09:35:00.0456029Z DOCS=yes 2025-12-04T09:35:00.0456179Z SHLVL=1 2025-12-04T09:35:00.0456321Z MAX_JOBS=30 2025-12-04T09:35:00.0456465Z GITHUB_ACTOR_ID=97764156 2025-12-04T09:35:00.0456690Z GITHUB_WORKFLOW_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:35:00.0456935Z GITHUB_REF_NAME=main 2025-12-04T09:35:00.0457174Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:35:00.0457439Z GITHUB_JOB=test 2025-12-04T09:35:00.0457600Z NO_TEST_TIMEOUT=False 2025-12-04T09:35:00.0457759Z TD_DISTRIBUTED=False 2025-12-04T09:35:00.0457940Z GITHUB_REPOSITORY=pytorch/pytorch 2025-12-04T09:35:00.0458140Z GITHUB_RETENTION_DAYS=90 2025-12-04T09:35:00.0458310Z OPENSSL_DIR=/opt/openssl 2025-12-04T09:35:00.0458488Z GITHUB_ACTION_REPOSITORY= 2025-12-04T09:35:00.0458947Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:35:00.0459402Z GITHUB_BASE_REF= 2025-12-04T09:35:00.0459555Z INSTALLED_ACL= 2025-12-04T09:35:00.0459857Z ARTIFACTS_FILE_SUFFIX=test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344 2025-12-04T09:35:00.0460179Z CI=true 2025-12-04T09:35:00.0460328Z GITHUB_REPOSITORY_OWNER=pytorch 2025-12-04T09:35:00.0460636Z RUST_LOG=sccache::server=error 2025-12-04T09:35:00.0460814Z JOB_ID=57118563344 2025-12-04T09:35:00.0460958Z GITHUB_HEAD_REF= 2025-12-04T09:35:00.0461112Z GITHUB_ACTION_REF= 2025-12-04T09:35:00.0461306Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-12-04T09:35:00.0461518Z TEST_SHOWLOCALS=False 2025-12-04T09:35:00.0461702Z GITHUB_WORKFLOW=inductor-periodic 2025-12-04T09:35:00.0461903Z DEBIAN_FRONTEND=noninteractive 2025-12-04T09:35:00.0462265Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_53a1234b-4002-441a-a1d2-f0fc0a79a49e 2025-12-04T09:35:00.0462614Z NO_TD=False 2025-12-04T09:35:00.0462773Z SKIP_SCCACHE_INITIALIZATION=1 2025-12-04T09:35:00.0462973Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-12-04T09:35:00.0463165Z _=/usr/bin/env 2025-12-04T09:35:00.0463381Z ++ python -c 'import site; print(site.getsitepackages()[0])' 2025-12-04T09:35:00.0692024Z + TORCH_INSTALL_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch 2025-12-04T09:35:00.0696812Z + TORCH_BIN_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:35:00.0697298Z + TORCH_LIB_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib 2025-12-04T09:35:00.0702322Z + TORCH_TEST_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/test 2025-12-04T09:35:00.0706897Z + BUILD_DIR=build 2025-12-04T09:35:00.0712000Z + BUILD_RENAMED_DIR=build_renamed 2025-12-04T09:35:00.0716625Z + BUILD_BIN_DIR=build/bin 2025-12-04T09:35:00.0718417Z + SHARD_NUMBER=1 2025-12-04T09:35:00.0718590Z + NUM_TEST_SHARDS=1 2025-12-04T09:35:00.0718795Z + export TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:35:00.0719022Z + TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:35:00.0719227Z + export VALGRIND=ON 2025-12-04T09:35:00.0719401Z + VALGRIND=ON 2025-12-04T09:35:00.0719601Z + [[ linux-jammy-py3.10-gcc11-build == *clang9* ]] 2025-12-04T09:35:00.0719855Z + [[ linux-jammy-py3.10-gcc11-build == *xpu* ]] 2025-12-04T09:35:00.0720068Z + detect_cuda_arch 2025-12-04T09:35:00.0720262Z + [[ linux-jammy-py3.10-gcc11-build == *cuda* ]] 2025-12-04T09:35:00.0720542Z + [[ linux-jammy-py3.10-gcc11-build == *s390x* ]] 2025-12-04T09:35:00.0720754Z + [[ 0 == \1 ]] 2025-12-04T09:35:00.0720913Z + [[ True == \1 ]] 2025-12-04T09:35:00.0721106Z + [[ linux-jammy-py3.10-gcc11-build != *bazel* ]] 2025-12-04T09:35:00.0721579Z ++ realpath build/custom_test_artifacts 2025-12-04T09:35:00.0722051Z + CUSTOM_TEST_ARTIFACT_BUILD_DIR=/var/lib/jenkins/workspace/build/custom_test_artifacts 2025-12-04T09:35:00.0722378Z + [[ -n '' ]] 2025-12-04T09:35:00.0722560Z + echo 'Environment variables' 2025-12-04T09:35:00.0722766Z Environment variables 2025-12-04T09:35:00.0722956Z + env 2025-12-04T09:35:00.0741379Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:35:00.0743137Z CONTINUE_THROUGH_ERROR=True 2025-12-04T09:35:00.0743617Z BUILD_ENVIRONMENT=linux-jammy-py3.10-gcc11-build 2025-12-04T09:35:00.0744202Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-12-04T09:35:00.0749238Z HOSTNAME=25f649d00d6e 2025-12-04T09:35:00.0749723Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_53a1234b-4002-441a-a1d2-f0fc0a79a49e 2025-12-04T09:35:00.0750111Z GITHUB_ACTION=__run_3 2025-12-04T09:35:00.0750303Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-12-04T09:35:00.0750502Z GITHUB_RUN_NUMBER=67027 2025-12-04T09:35:00.0750728Z TEST_CONFIG=cpu_inductor_amp_freezing_huggingface 2025-12-04T09:35:00.0750970Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-12-04T09:35:00.0751193Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-12-04T09:35:00.0751405Z SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:35:00.0751762Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-12-04T09:35:00.0751968Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-12-04T09:35:00.0752171Z GITHUB_REF_TYPE=branch 2025-12-04T09:35:00.0752374Z BASE_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:35:00.0752591Z XLA_CUDA= 2025-12-04T09:35:00.0752741Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-12-04T09:35:00.0753166Z HUGGING_FACE_HUB_TOKEN=*** 2025-12-04T09:35:00.0753410Z *** 2025-12-04T09:35:00.0753561Z GITHUB_REPOSITORY_ID=65600975 2025-12-04T09:35:00.0753735Z GITHUB_ACTIONS=true 2025-12-04T09:35:00.0753929Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:35:00.0754175Z SHA1=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:35:00.0754414Z GITHUB_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:35:00.0754764Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/inductor-periodic.yml@refs/heads/main 2025-12-04T09:35:00.0755075Z UCC_HOME=/usr 2025-12-04T09:35:00.0755237Z TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:35:00.0755408Z VERBOSE_TEST_LOGS=False 2025-12-04T09:35:00.0755579Z GITHUB_REF=refs/heads/main 2025-12-04T09:35:00.0755752Z SHARD_NUMBER=1 2025-12-04T09:35:00.0755905Z GITHUB_REF_PROTECTED=true 2025-12-04T09:35:00.0756077Z HOME=/var/lib/jenkins 2025-12-04T09:35:00.0756263Z GITHUB_API_URL=https://api.github.com 2025-12-04T09:35:00.0756466Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-12-04T09:35:00.0756652Z UCX_COMMIT= 2025-12-04T09:35:00.0756800Z USE_SYSTEM_NCCL=1 2025-12-04T09:35:00.0756947Z NUM_TEST_SHARDS=1 2025-12-04T09:35:00.0757099Z UCX_HOME=/usr 2025-12-04T09:35:00.0757431Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_53a1234b-4002-441a-a1d2-f0fc0a79a49e 2025-12-04T09:35:00.0757970Z JOB_NAME=periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx) 2025-12-04T09:35:00.0758498Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_53a1234b-4002-441a-a1d2-f0fc0a79a49e 2025-12-04T09:35:00.0758952Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-12-04T09:35:00.0759246Z GITHUB_EVENT_NAME=schedule 2025-12-04T09:35:00.0759413Z DASHBOARD_TAG= 2025-12-04T09:35:00.0759569Z GITHUB_RUN_ID=19923066595 2025-12-04T09:35:00.0759740Z INSTALLED_OPENBLAS= 2025-12-04T09:35:00.0760086Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_53a1234b-4002-441a-a1d2-f0fc0a79a49e 2025-12-04T09:35:00.0760473Z GITHUB_ACTOR=pytorchmergebot 2025-12-04T09:35:00.0760655Z PR_NUMBER= 2025-12-04T09:35:00.0760790Z DESIRED_CUDA= 2025-12-04T09:35:00.0760942Z GITHUB_RUN_ATTEMPT=1 2025-12-04T09:35:00.0761103Z VALGRIND=ON 2025-12-04T09:35:00.0761257Z ANACONDA_PYTHON_VERSION=3.10 2025-12-04T09:35:00.0761665Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-12-04T09:35:00.0762038Z TERM=vt100 2025-12-04T09:35:00.0762191Z INSTALLED_VISION=yes 2025-12-04T09:35:00.0762349Z BRANCH=main 2025-12-04T09:35:00.0762507Z SCCACHE_REGION=us-east-1 2025-12-04T09:35:00.0762699Z OPENSSL_ROOT_DIR=/opt/openssl 2025-12-04T09:35:00.0762886Z BUILD_AOT_INDUCTOR_TEST=yes 2025-12-04T09:35:00.0763156Z CUDA_PATH=/usr/local/cuda 2025-12-04T09:35:00.0763468Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-12-04T09:35:00.0763822Z GITHUB_SERVER_URL=https://github.com 2025-12-04T09:35:00.0764020Z UCC_COMMIT= 2025-12-04T09:35:00.0764166Z REENABLED_ISSUES= 2025-12-04T09:35:00.0764316Z DOCS=yes 2025-12-04T09:35:00.0764459Z SHLVL=1 2025-12-04T09:35:00.0764601Z MAX_JOBS=30 2025-12-04T09:35:00.0764743Z GITHUB_ACTOR_ID=97764156 2025-12-04T09:35:00.0764971Z GITHUB_WORKFLOW_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:35:00.0765213Z GITHUB_REF_NAME=main 2025-12-04T09:35:00.0765464Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:35:00.0765722Z GITHUB_JOB=test 2025-12-04T09:35:00.0765881Z NO_TEST_TIMEOUT=False 2025-12-04T09:35:00.0766046Z TD_DISTRIBUTED=False 2025-12-04T09:35:00.0766215Z GITHUB_REPOSITORY=pytorch/pytorch 2025-12-04T09:35:00.0766410Z GITHUB_RETENTION_DAYS=90 2025-12-04T09:35:00.0766590Z OPENSSL_DIR=/opt/openssl 2025-12-04T09:35:00.0766760Z GITHUB_ACTION_REPOSITORY= 2025-12-04T09:35:00.0767223Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:35:00.0767683Z GITHUB_BASE_REF= 2025-12-04T09:35:00.0767835Z INSTALLED_ACL= 2025-12-04T09:35:00.0768139Z ARTIFACTS_FILE_SUFFIX=test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344 2025-12-04T09:35:00.0768461Z CI=true 2025-12-04T09:35:00.0768616Z GITHUB_REPOSITORY_OWNER=pytorch 2025-12-04T09:35:00.0768845Z RUST_LOG=sccache::server=error 2025-12-04T09:35:00.0769025Z JOB_ID=57118563344 2025-12-04T09:35:00.0769181Z GITHUB_HEAD_REF= 2025-12-04T09:35:00.0769329Z GITHUB_ACTION_REF= 2025-12-04T09:35:00.0769525Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-12-04T09:35:00.0769749Z TEST_SHOWLOCALS=False 2025-12-04T09:35:00.0769922Z GITHUB_WORKFLOW=inductor-periodic 2025-12-04T09:35:00.0770126Z DEBIAN_FRONTEND=noninteractive 2025-12-04T09:35:00.0770500Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_53a1234b-4002-441a-a1d2-f0fc0a79a49e 2025-12-04T09:35:00.0770861Z NO_TD=False 2025-12-04T09:35:00.0771021Z SKIP_SCCACHE_INITIALIZATION=1 2025-12-04T09:35:00.0771226Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-12-04T09:35:00.0771419Z _=/usr/bin/env 2025-12-04T09:35:00.0771579Z + echo 'Testing pytorch' 2025-12-04T09:35:00.0771750Z Testing pytorch 2025-12-04T09:35:00.0771914Z + export LANG=C.UTF-8 2025-12-04T09:35:00.0772079Z + LANG=C.UTF-8 2025-12-04T09:35:00.0772230Z + PR_NUMBER= 2025-12-04T09:35:00.0772437Z + [[ cpu_inductor_amp_freezing_huggingface == \d\e\f\a\u\l\t ]] 2025-12-04T09:35:00.0772735Z + [[ cpu_inductor_amp_freezing_huggingface == \d\i\s\t\r\i\b\u\t\e\d ]] 2025-12-04T09:35:00.0773022Z + [[ cpu_inductor_amp_freezing_huggingface == \s\l\o\w ]] 2025-12-04T09:35:00.0773296Z + [[ linux-jammy-py3.10-gcc11-build == *slow-gradcheck* ]] 2025-12-04T09:35:00.0773556Z + [[ linux-jammy-py3.10-gcc11-build == *cuda* ]] 2025-12-04T09:35:00.0773791Z + [[ linux-jammy-py3.10-gcc11-build == *rocm* ]] 2025-12-04T09:35:00.0774023Z + [[ linux-jammy-py3.10-gcc11-build == *xpu* ]] 2025-12-04T09:35:00.0774272Z + [[ cpu_inductor_amp_freezing_huggingface == *crossref* ]] 2025-12-04T09:35:00.0774522Z + [[ linux-jammy-py3.10-gcc11-build == *rocm* ]] 2025-12-04T09:35:00.0774752Z + [[ linux-jammy-py3.10-gcc11-build == *xpu* ]] 2025-12-04T09:35:00.0774990Z + [[ linux-jammy-py3.10-gcc11-build != *-bazel-* ]] 2025-12-04T09:35:00.0775208Z + pip_install ninja==1.10.2 2025-12-04T09:35:00.0775485Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-12-04T09:35:00.0775783Z + python3 -m pip install --progress-bar off ninja==1.10.2 2025-12-04T09:35:00.4040797Z Collecting ninja==1.10.2 2025-12-04T09:35:00.4248406Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl.metadata (5.0 kB) 2025-12-04T09:35:00.4371724Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl (108 kB) 2025-12-04T09:35:01.1579754Z Installing collected packages: ninja 2025-12-04T09:35:01.1580081Z Attempting uninstall: ninja 2025-12-04T09:35:01.1591186Z Found existing installation: ninja 1.11.1.4 2025-12-04T09:35:01.1602784Z Uninstalling ninja-1.11.1.4: 2025-12-04T09:35:01.1653665Z Successfully uninstalled ninja-1.11.1.4 2025-12-04T09:35:01.2106078Z Successfully installed ninja-1.10.2 2025-12-04T09:35:01.3034134Z + export PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:35:01.3035169Z + PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:35:01.3035720Z + [[ linux-jammy-py3.10-gcc11-build == *aarch64* ]] 2025-12-04T09:35:01.3035989Z + [[ linux-jammy-py3.10-gcc11-build == *asan* ]] 2025-12-04T09:35:01.3036230Z + [[ linux-jammy-py3.10-gcc11-build == *-debug* ]] 2025-12-04T09:35:01.3036475Z + [[ linux-jammy-py3.10-gcc11-build != *-bazel-* ]] 2025-12-04T09:35:01.3036807Z + echo 'We are not in debug mode: linux-jammy-py3.10-gcc11-build. Expect the assertion to pass' 2025-12-04T09:35:01.3037193Z We are not in debug mode: linux-jammy-py3.10-gcc11-build. Expect the assertion to pass 2025-12-04T09:35:01.3037468Z + cd test 2025-12-04T09:35:01.3037696Z + python -c 'import torch; torch._C._crash_if_debug_asserts_fail(424242)' 2025-12-04T09:35:01.5645665Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:35:01.5646562Z import pynvml # type: ignore[import] 2025-12-04T09:35:02.4075201Z + [[ cpu_inductor_amp_freezing_huggingface == \n\o\g\p\u\_\N\O\_\A\V\X\2 ]] 2025-12-04T09:35:02.4077456Z + [[ cpu_inductor_amp_freezing_huggingface == \n\o\g\p\u\_\A\V\X\5\1\2 ]] 2025-12-04T09:35:02.4078035Z + [[ cpu_inductor_amp_freezing_huggingface == \l\e\g\a\c\y\_\n\v\i\d\i\a\_\d\r\i\v\e\r ]] 2025-12-04T09:35:02.4078533Z + DYNAMO_BENCHMARK_FLAGS=() 2025-12-04T09:35:02.4078947Z + [[ cpu_inductor_amp_freezing_huggingface == *pr_time_benchmarks* ]] 2025-12-04T09:35:02.4079293Z + [[ cpu_inductor_amp_freezing_huggingface == *dynamo_eager* ]] 2025-12-04T09:35:02.4079634Z + [[ cpu_inductor_amp_freezing_huggingface == *aot_eager* ]] 2025-12-04T09:35:02.4079950Z + [[ cpu_inductor_amp_freezing_huggingface == *aot_inductor* ]] 2025-12-04T09:35:02.4080287Z + [[ cpu_inductor_amp_freezing_huggingface == *max_autotune_inductor* ]] 2025-12-04T09:35:02.4080616Z + [[ cpu_inductor_amp_freezing_huggingface == *inductor* ]] 2025-12-04T09:35:02.4080905Z + [[ cpu_inductor_amp_freezing_huggingface != *perf* ]] 2025-12-04T09:35:02.4081191Z + DYNAMO_BENCHMARK_FLAGS+=(--inductor) 2025-12-04T09:35:02.4081451Z + [[ cpu_inductor_amp_freezing_huggingface == *dynamic* ]] 2025-12-04T09:35:02.4081708Z + [[ cpu_inductor_amp_freezing_huggingface == *cpu* ]] 2025-12-04T09:35:02.4082093Z + DYNAMO_BENCHMARK_FLAGS+=(--device cpu) 2025-12-04T09:35:02.4228628Z + [[ linux-jammy-py3.10-gcc11-build == *libtorch* ]] 2025-12-04T09:35:02.4228939Z + [[ linux-jammy-py3.10-gcc11-build == *-bazel-* ]] 2025-12-04T09:35:02.4229161Z + cd test 2025-12-04T09:35:02.4229370Z + python -c 'import torch; print(torch.__config__.show())' 2025-12-04T09:35:02.6823809Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:35:02.6825332Z import pynvml # type: ignore[import] 2025-12-04T09:35:03.2816117Z PyTorch built with: 2025-12-04T09:35:03.2821946Z - GCC 11.4 2025-12-04T09:35:03.2822314Z - C++ Version: 201703 2025-12-04T09:35:03.2822777Z - Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-12-04T09:35:03.2823381Z - Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-12-04T09:35:03.2823668Z - OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-12-04T09:35:03.2823903Z - LAPACK is enabled (usually provided by MKL) 2025-12-04T09:35:03.2824112Z - NNPACK is enabled 2025-12-04T09:35:03.2824296Z - CPU capability usage: AVX512 2025-12-04T09:35:03.2826881Z - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, COMMIT_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32, CXX_COMPILER=/opt/cache/bin/c++, CXX_FLAGS= -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOCUPTI -DLIBKINETO_NOROCTRACER -DLIBKINETO_NOXPUPTI=ON -DUSE_FBGEMM -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -DC10_NODEPRECATED -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-unknown-pragmas -Wno-unused-parameter -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=old-style-cast -faligned-new -Werror -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, TORCH_VERSION=2.10.0, USE_CUDA=OFF, USE_CUDNN=OFF, USE_CUSPARSELT=OFF, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=OFF, USE_NCCL=OFF, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, USE_ROCM_KERNEL_ASSERT=OFF, USE_XCCL=OFF, USE_XPU=OFF, 2025-12-04T09:35:03.2829465Z 2025-12-04T09:35:03.4780436Z + cd test 2025-12-04T09:35:03.4782565Z + python -c 'import torch; print(torch.__config__.parallel_info())' 2025-12-04T09:35:03.7335497Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:35:03.7336433Z import pynvml # type: ignore[import] 2025-12-04T09:35:04.3385390Z ATen/Parallel: 2025-12-04T09:35:04.3387049Z at::get_num_threads() : 16 2025-12-04T09:35:04.3387438Z at::get_num_interop_threads() : 16 2025-12-04T09:35:04.3389818Z OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-12-04T09:35:04.3390169Z omp_get_max_threads() : 16 2025-12-04T09:35:04.3395194Z Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-12-04T09:35:04.3397208Z mkl_get_max_threads() : 16 2025-12-04T09:35:04.3397537Z Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-12-04T09:35:04.3397849Z std::thread::hardware_concurrency() : 32 2025-12-04T09:35:04.3398051Z Environment variables: 2025-12-04T09:35:04.3398267Z OMP_NUM_THREADS : [not set] 2025-12-04T09:35:04.3398448Z MKL_NUM_THREADS : [not set] 2025-12-04T09:35:04.3398621Z ATen parallel backend: OpenMP 2025-12-04T09:35:04.3398748Z 2025-12-04T09:35:04.5328422Z + [[ cpu_inductor_amp_freezing_huggingface == *numpy_2* ]] 2025-12-04T09:35:04.5328780Z + [[ linux-jammy-py3.10-gcc11-build == *aarch64* ]] 2025-12-04T09:35:04.5329062Z + [[ cpu_inductor_amp_freezing_huggingface == *backward* ]] 2025-12-04T09:35:04.5329434Z + [[ cpu_inductor_amp_freezing_huggingface == *libtorch_agnostic_targetting* ]] 2025-12-04T09:35:04.5329758Z + [[ cpu_inductor_amp_freezing_huggingface == *xla* ]] 2025-12-04T09:35:04.5330537Z + [[ cpu_inductor_amp_freezing_huggingface == *vllm* ]] 2025-12-04T09:35:04.5330851Z + [[ cpu_inductor_amp_freezing_huggingface == *executorch* ]] 2025-12-04T09:35:04.5331155Z + [[ cpu_inductor_amp_freezing_huggingface == \j\i\t\_\l\e\g\a\c\y ]] 2025-12-04T09:35:04.5331480Z + [[ cpu_inductor_amp_freezing_huggingface == \q\u\a\n\t\i\z\a\t\i\o\n ]] 2025-12-04T09:35:04.5331898Z + [[ linux-jammy-py3.10-gcc11-build == *libtorch* ]] 2025-12-04T09:35:04.5332177Z + [[ cpu_inductor_amp_freezing_huggingface == distributed ]] 2025-12-04T09:35:04.5332484Z + [[ cpu_inductor_amp_freezing_huggingface == *operator_benchmark* ]] 2025-12-04T09:35:04.5332842Z + [[ cpu_inductor_amp_freezing_huggingface == *operator_microbenchmark* ]] 2025-12-04T09:35:04.5333185Z + [[ cpu_inductor_amp_freezing_huggingface == *attention_microbenchmark* ]] 2025-12-04T09:35:04.5333546Z + [[ cpu_inductor_amp_freezing_huggingface == *inductor_distributed* ]] 2025-12-04T09:35:04.5333894Z + [[ cpu_inductor_amp_freezing_huggingface == *inductor-halide* ]] 2025-12-04T09:35:04.5334215Z + [[ cpu_inductor_amp_freezing_huggingface == *inductor-pallas* ]] 2025-12-04T09:35:04.5334530Z + [[ cpu_inductor_amp_freezing_huggingface == *inductor-triton-cpu* ]] 2025-12-04T09:35:04.5334868Z + [[ cpu_inductor_amp_freezing_huggingface == *inductor-micro-benchmark* ]] 2025-12-04T09:35:04.5335222Z + [[ cpu_inductor_amp_freezing_huggingface == *aoti_cross_compile_for_windows* ]] 2025-12-04T09:35:04.5335558Z + [[ cpu_inductor_amp_freezing_huggingface == *huggingface* ]] 2025-12-04T09:35:04.5335808Z + install_torchvision 2025-12-04T09:35:04.5335995Z + local orig_preload 2025-12-04T09:35:04.5336152Z + local commit 2025-12-04T09:35:04.5336398Z ++ get_pinned_commit vision 2025-12-04T09:35:04.5338227Z ++ cat .github/ci_commit_pins/vision.txt 2025-12-04T09:35:04.5350541Z + commit=617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:35:04.5354712Z + orig_preload= 2025-12-04T09:35:04.5358667Z + '[' -n '' ']' 2025-12-04T09:35:04.5358926Z + [[ linux-jammy-py3.10-gcc11-build == *cuda* ]] 2025-12-04T09:35:04.5359379Z + pip_build_and_install git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e dist/vision 2025-12-04T09:35:04.5359896Z + local build_target=git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:35:04.5376359Z + local wheel_dir=dist/vision 2025-12-04T09:35:04.5376675Z + local found_whl=0 2025-12-04T09:35:04.5376856Z + for file in "${wheel_dir}"/*.whl 2025-12-04T09:35:04.5377173Z + [[ -f dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl ]] 2025-12-04T09:35:04.5377477Z + found_whl=1 2025-12-04T09:35:04.5377629Z + break 2025-12-04T09:35:04.5377763Z + '[' 1 == 0 ']' 2025-12-04T09:35:04.5377933Z + for file in "${wheel_dir}"/*.whl 2025-12-04T09:35:04.5378241Z + pip_install_whl dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:35:04.5378630Z + args=('dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl') 2025-12-04T09:35:04.5378909Z + local args 2025-12-04T09:35:04.5379167Z + [[ dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl == *\ * ]] 2025-12-04T09:35:04.5379456Z + for path in "${args[@]}" 2025-12-04T09:35:04.5379746Z + echo 'Installing dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl' 2025-12-04T09:35:04.5380136Z Installing dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:35:04.5380581Z + python3 -mpip install --no-index --no-deps dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:35:04.7985246Z Processing ./dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:35:04.8056522Z Installing collected packages: torchvision 2025-12-04T09:35:05.4074133Z Successfully installed torchvision-0.25.0a0+617079d 2025-12-04T09:35:05.4479764Z + '[' -n '' ']' 2025-12-04T09:35:05.4480001Z + id=0 2025-12-04T09:35:05.4480188Z + test_dynamo_benchmark huggingface 0 2025-12-04T09:35:05.4480406Z ++ pwd 2025-12-04T09:35:05.4486387Z + TEST_REPORTS_DIR=/var/lib/jenkins/workspace/test/test-reports 2025-12-04T09:35:05.4490541Z + local suite=huggingface 2025-12-04T09:35:05.4492510Z + shift 2025-12-04T09:35:05.4492828Z + local shard_id=0 2025-12-04T09:35:05.4497213Z + shift 2025-12-04T09:35:05.4497490Z + extra_args=() 2025-12-04T09:35:05.4501418Z + local extra_args 2025-12-04T09:35:05.4501970Z + [[ linux-jammy-py3.10-gcc11-build == *cuda13* ]] 2025-12-04T09:35:05.4502265Z + [[ cpu_inductor_amp_freezing_huggingface == *perf_compare* ]] 2025-12-04T09:35:05.4502540Z + [[ cpu_inductor_amp_freezing_huggingface == *perf* ]] 2025-12-04T09:35:05.4502800Z + [[ cpu_inductor_amp_freezing_huggingface == *cpu* ]] 2025-12-04T09:35:05.4503014Z + local dt=float32 2025-12-04T09:35:05.4503196Z + [[ cpu_inductor_amp_freezing_huggingface == *amp* ]] 2025-12-04T09:35:05.4503404Z + dt=amp 2025-12-04T09:35:05.4503591Z + [[ cpu_inductor_amp_freezing_huggingface == *freezing* ]] 2025-12-04T09:35:05.4503917Z + test_single_dynamo_benchmark inference huggingface 0 --inference --amp --freezing 2025-12-04T09:35:05.4504208Z ++ pwd 2025-12-04T09:35:05.4504423Z + TEST_REPORTS_DIR=/var/lib/jenkins/workspace/test/test-reports 2025-12-04T09:35:05.4504701Z + mkdir -p /var/lib/jenkins/workspace/test/test-reports 2025-12-04T09:35:05.4507674Z + local name=inference 2025-12-04T09:35:05.4513340Z + shift 2025-12-04T09:35:05.4513590Z + local suite=huggingface 2025-12-04T09:35:05.4513825Z + shift 2025-12-04T09:35:05.4514008Z + local shard_id=0 2025-12-04T09:35:05.4514204Z + shift 2025-12-04T09:35:05.4514351Z + partition_flags=() 2025-12-04T09:35:05.4514541Z + local partition_flags 2025-12-04T09:35:05.4514758Z + [[ -n 1 ]] 2025-12-04T09:35:05.4514949Z + [[ -n 0 ]] 2025-12-04T09:35:05.4515250Z + partition_flags=(--total-partitions "$NUM_TEST_SHARDS" --partition-id "$shard_id") 2025-12-04T09:35:05.4515645Z + [[ cpu_inductor_amp_freezing_huggingface == *perf_compare* ]] 2025-12-04T09:35:05.4515919Z + [[ cpu_inductor_amp_freezing_huggingface == *perf* ]] 2025-12-04T09:35:05.4516202Z + [[ cpu_inductor_amp_freezing_huggingface == *_avx2* ]] 2025-12-04T09:35:05.4516467Z + [[ cpu_inductor_amp_freezing_huggingface == *_avx512* ]] 2025-12-04T09:35:05.4517222Z + python benchmarks/dynamo/huggingface.py --ci --accuracy --timing --explain --print-compilation-time --inductor --device cpu --inference --amp --freezing --total-partitions 1 --partition-id 0 --output /var/lib/jenkins/workspace/test/test-reports/inference_huggingface.csv 2025-12-04T09:35:06.2029680Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:35:06.2030780Z import pynvml # type: ignore[import] 2025-12-04T09:35:09.2357666Z 2025-12-04T09:35:09.2358284Z config.json: 0% 0.00/694 [00:00bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8143541Z 2025-12-04T09:36:29.8143620Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8143829Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8144066Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8144588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8145083Z layer_outputs = layer_module( 2025-12-04T09:36:29.8145440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8145812Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8146284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8146705Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8147112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8147518Z self_outputs = self.self( 2025-12-04T09:36:29.8147952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8148390Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8148873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8149447Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8149687Z 2025-12-04T09:36:29.8149802Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8150315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8150793Z layer_outputs = layer_module( 2025-12-04T09:36:29.8151151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8151530Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8151942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8152355Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8152764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8153221Z self_outputs = self.self( 2025-12-04T09:36:29.8153619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8154071Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8154580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8155236Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8155485Z 2025-12-04T09:36:29.8155593Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8156117Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8156632Z layer_outputs = layer_module( 2025-12-04T09:36:29.8156999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8157372Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8157819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8158271Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8158728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8159164Z self_outputs = self.self( 2025-12-04T09:36:29.8159628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8160097Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8160634Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8161331Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8161688Z 2025-12-04T09:36:29.8161792Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8162033Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8162267Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8162540Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8162790Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8163344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8163834Z layer_outputs = layer_module( 2025-12-04T09:36:29.8164189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8164552Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8164982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8165404Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8165816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8166234Z self_outputs = self.self( 2025-12-04T09:36:29.8166640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 535, in forward 2025-12-04T09:36:29.8167099Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8167636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 833, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8168205Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-12-04T09:36:29.8168741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 763, in _mask_invalid_locations 2025-12-04T09:36:29.8169249Z ).where(beginning_mask.bool(), beginning_input) 2025-12-04T09:36:29.8169412Z 2025-12-04T09:36:29.8169497Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8169754Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8170310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8170830Z layer_outputs = layer_module( 2025-12-04T09:36:29.8171188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8171557Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8171995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8172435Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8172886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8173328Z self_outputs = self.self( 2025-12-04T09:36:29.8173756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 540, in forward 2025-12-04T09:36:29.8174193Z attn_scores += diagonal_mask 2025-12-04T09:36:29.8174334Z 2025-12-04T09:36:29.8174447Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8174984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8175475Z layer_outputs = layer_module( 2025-12-04T09:36:29.8175867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8176261Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8176725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8177169Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8177650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8178090Z self_outputs = self.self( 2025-12-04T09:36:29.8178517Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 578, in forward 2025-12-04T09:36:29.8178958Z attn_probs = nn.functional.softmax( 2025-12-04T09:36:29.8179108Z 2025-12-04T09:36:29.8179193Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8179420Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8179678Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8180221Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8180738Z layer_outputs = layer_module( 2025-12-04T09:36:29.8181116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8181499Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8181953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8182396Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8182839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8183269Z self_outputs = self.self( 2025-12-04T09:36:29.8183693Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8184158Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8184702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 862, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8185328Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-12-04T09:36:29.8185788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8186175Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8186336Z 2025-12-04T09:36:29.8186456Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8186999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8187514Z layer_outputs = layer_module( 2025-12-04T09:36:29.8187885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8188269Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8188719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8189166Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8189612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8190043Z self_outputs = self.self( 2025-12-04T09:36:29.8190467Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8190999Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8191581Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 875, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8192172Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-12-04T09:36:29.8192765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 698, in _pad_and_diagonalize 2025-12-04T09:36:29.8193276Z chunked_hidden_states = nn.functional.pad( 2025-12-04T09:36:29.8193650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8194027Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8194200Z 2025-12-04T09:36:29.8194314Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8194875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8195398Z layer_outputs = layer_module( 2025-12-04T09:36:29.8195776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8196174Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8196636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8197081Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8197509Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8197930Z self_outputs = self.self( 2025-12-04T09:36:29.8198339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8198807Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8199372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8199984Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8200211Z 2025-12-04T09:36:29.8200331Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8200887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8201414Z layer_outputs = layer_module( 2025-12-04T09:36:29.8201879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8202285Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8202754Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8203234Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8203698Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8204122Z self_outputs = self.self( 2025-12-04T09:36:29.8204571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8205065Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8205630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8206229Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8206503Z 2025-12-04T09:36:29.8206618Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8207172Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8207687Z layer_outputs = layer_module( 2025-12-04T09:36:29.8208085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8208469Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8208904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8209333Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8209767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8210190Z self_outputs = self.self( 2025-12-04T09:36:29.8210583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 617, in forward 2025-12-04T09:36:29.8211101Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-12-04T09:36:29.8211344Z 2025-12-04T09:36:29.8211425Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8211645Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8211857Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8212060Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8212264Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8212471Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8212694Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8213205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8213686Z layer_outputs = layer_module( 2025-12-04T09:36:29.8214031Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8214386Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8214811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8215235Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8215653Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8216068Z self_outputs = self.self( 2025-12-04T09:36:29.8216481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 508, in forward 2025-12-04T09:36:29.8216911Z query_vectors = self.query(hidden_states) 2025-12-04T09:36:29.8217054Z 2025-12-04T09:36:29.8217137Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8217350Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8217589Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8218098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8218595Z layer_outputs = layer_module( 2025-12-04T09:36:29.8218944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8219314Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8219734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8220155Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8220613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8221034Z self_outputs = self.self( 2025-12-04T09:36:29.8221427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8221873Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8222419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8223019Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8223268Z 2025-12-04T09:36:29.8223351Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8223570Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8223815Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8224339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8224846Z layer_outputs = layer_module( 2025-12-04T09:36:29.8225197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8225566Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8225982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8226403Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8226820Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8227234Z self_outputs = self.self( 2025-12-04T09:36:29.8227633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8228077Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8228575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8229151Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8229403Z 2025-12-04T09:36:29.8229512Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8230026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8230703Z layer_outputs = layer_module( 2025-12-04T09:36:29.8231058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8231439Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8231882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8232316Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8232739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8233180Z self_outputs = self.self( 2025-12-04T09:36:29.8233595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8234034Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8234528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8235229Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8235478Z 2025-12-04T09:36:29.8235596Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8236121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8236616Z layer_outputs = layer_module( 2025-12-04T09:36:29.8237024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8237393Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8238112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8238536Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8238964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8239388Z self_outputs = self.self( 2025-12-04T09:36:29.8239785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8240235Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8240742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8241339Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8241647Z 2025-12-04T09:36:29.8241742Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8241975Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8242230Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8242779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8243327Z layer_outputs = layer_module( 2025-12-04T09:36:29.8243691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8244087Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8244531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8244968Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8245413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8245854Z self_outputs = self.self( 2025-12-04T09:36:29.8246272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 540, in forward 2025-12-04T09:36:29.8246706Z attn_scores += diagonal_mask 2025-12-04T09:36:29.8246836Z 2025-12-04T09:36:29.8246956Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8247489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8247998Z layer_outputs = layer_module( 2025-12-04T09:36:29.8248365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8248767Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8249201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8249618Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8250022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8250458Z self_outputs = self.self( 2025-12-04T09:36:29.8250841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 578, in forward 2025-12-04T09:36:29.8251246Z attn_probs = nn.functional.softmax( 2025-12-04T09:36:29.8251375Z 2025-12-04T09:36:29.8251461Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8251695Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8251926Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8252419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8252887Z layer_outputs = layer_module( 2025-12-04T09:36:29.8253214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8253565Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8253972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8254373Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8254764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8255165Z self_outputs = self.self( 2025-12-04T09:36:29.8255547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8255978Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8256488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 862, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8257053Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-12-04T09:36:29.8257473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8257808Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8257961Z 2025-12-04T09:36:29.8258062Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8258558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8259027Z layer_outputs = layer_module( 2025-12-04T09:36:29.8259353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8259703Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8260106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8260514Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8260906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8261301Z self_outputs = self.self( 2025-12-04T09:36:29.8261682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8262115Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8262618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 875, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8263137Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-12-04T09:36:29.8263626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 698, in _pad_and_diagonalize 2025-12-04T09:36:29.8264102Z chunked_hidden_states = nn.functional.pad( 2025-12-04T09:36:29.8264436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8264780Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8264926Z 2025-12-04T09:36:29.8265035Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8265554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8266022Z layer_outputs = layer_module( 2025-12-04T09:36:29.8266360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8266712Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8267111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8267521Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8267912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8268298Z self_outputs = self.self( 2025-12-04T09:36:29.8268673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8269100Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8269587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8270107Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8270305Z 2025-12-04T09:36:29.8270404Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8270888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8271340Z layer_outputs = layer_module( 2025-12-04T09:36:29.8271662Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8272004Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8272403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8272791Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8273172Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8273560Z self_outputs = self.self( 2025-12-04T09:36:29.8273947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8274376Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8274877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8275421Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8275612Z 2025-12-04T09:36:29.8275721Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8276205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8276668Z layer_outputs = layer_module( 2025-12-04T09:36:29.8277001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8277382Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8277779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8278183Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8278581Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8279012Z self_outputs = self.self( 2025-12-04T09:36:29.8279386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 617, in forward 2025-12-04T09:36:29.8279890Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-12-04T09:36:29.8280118Z 2025-12-04T09:36:29.8280205Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8280413Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8280616Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8280821Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8281022Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8281219Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8281461Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8282112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8282659Z layer_outputs = layer_module( 2025-12-04T09:36:29.8283049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8283423Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8283857Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8284294Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8284778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8285203Z self_outputs = self.self( 2025-12-04T09:36:29.8285620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 508, in forward 2025-12-04T09:36:29.8286059Z query_vectors = self.query(hidden_states) 2025-12-04T09:36:29.8286211Z 2025-12-04T09:36:29.8286292Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8286507Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8286738Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8287262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8287754Z layer_outputs = layer_module( 2025-12-04T09:36:29.8288112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8288476Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8288900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8289318Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8289737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8290157Z self_outputs = self.self( 2025-12-04T09:36:29.8290561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8291031Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8291618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8292213Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8292465Z 2025-12-04T09:36:29.8292547Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8292763Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8292999Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8293558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8294061Z layer_outputs = layer_module( 2025-12-04T09:36:29.8294416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8294782Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8295211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8295635Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8296055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8296475Z self_outputs = self.self( 2025-12-04T09:36:29.8296877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8297328Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8297830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8298424Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8298678Z 2025-12-04T09:36:29.8298786Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8299314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8299795Z layer_outputs = layer_module( 2025-12-04T09:36:29.8300151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8300524Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8300961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8301365Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8301779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8302183Z self_outputs = self.self( 2025-12-04T09:36:29.8302569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8303011Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8303501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8304080Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8304322Z 2025-12-04T09:36:29.8304426Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8304934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8305413Z layer_outputs = layer_module( 2025-12-04T09:36:29.8305759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8306147Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8306567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8306979Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8307393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8307826Z self_outputs = self.self( 2025-12-04T09:36:29.8308219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8308655Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8309139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8309715Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8309969Z 2025-12-04T09:36:29.8310049Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8310262Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8310491Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8311004Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8311493Z layer_outputs = layer_module( 2025-12-04T09:36:29.8311826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8312174Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8312590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8313004Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8313397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8313791Z self_outputs = self.self( 2025-12-04T09:36:29.8314175Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 540, in forward 2025-12-04T09:36:29.8314580Z attn_scores += diagonal_mask 2025-12-04T09:36:29.8314696Z 2025-12-04T09:36:29.8314796Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8315288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8315753Z layer_outputs = layer_module( 2025-12-04T09:36:29.8316085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8316428Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8316831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8317231Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8317625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8318023Z self_outputs = self.self( 2025-12-04T09:36:29.8318414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 578, in forward 2025-12-04T09:36:29.8318824Z attn_probs = nn.functional.softmax( 2025-12-04T09:36:29.8318956Z 2025-12-04T09:36:29.8319033Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8319244Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8319478Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8320020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8320492Z layer_outputs = layer_module( 2025-12-04T09:36:29.8320835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8321237Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8321740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8322226Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8322690Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8323148Z self_outputs = self.self( 2025-12-04T09:36:29.8323561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8324024Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8324558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 862, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8325159Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-12-04T09:36:29.8325595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8325962Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8326119Z 2025-12-04T09:36:29.8326245Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8326759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8327235Z layer_outputs = layer_module( 2025-12-04T09:36:29.8327586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8327951Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8328363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8328782Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8329195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8329603Z self_outputs = self.self( 2025-12-04T09:36:29.8329990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8330608Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8331135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 875, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8331668Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-12-04T09:36:29.8332164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 698, in _pad_and_diagonalize 2025-12-04T09:36:29.8332631Z chunked_hidden_states = nn.functional.pad( 2025-12-04T09:36:29.8332974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8333318Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8333476Z 2025-12-04T09:36:29.8333582Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8334165Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8334650Z layer_outputs = layer_module( 2025-12-04T09:36:29.8334999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8335366Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8335790Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8336270Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8336678Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8337084Z self_outputs = self.self( 2025-12-04T09:36:29.8337474Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8337924Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8338433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8338986Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8339190Z 2025-12-04T09:36:29.8339302Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8339816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8340287Z layer_outputs = layer_module( 2025-12-04T09:36:29.8340639Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8341000Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8341413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8341828Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8342239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8342643Z self_outputs = self.self( 2025-12-04T09:36:29.8343030Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8343484Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8344001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8344557Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8344759Z 2025-12-04T09:36:29.8344862Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8345374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8345863Z layer_outputs = layer_module( 2025-12-04T09:36:29.8346206Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8346558Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8346965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8347372Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8347764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8348168Z self_outputs = self.self( 2025-12-04T09:36:29.8348588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 617, in forward 2025-12-04T09:36:29.8349098Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-12-04T09:36:29.8349328Z 2025-12-04T09:36:29.8349407Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8349618Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8349857Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8350050Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8350250Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8350449Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8350678Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8351168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8351639Z layer_outputs = layer_module( 2025-12-04T09:36:29.8351985Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8352333Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8352749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8353165Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8353567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8353955Z self_outputs = self.self( 2025-12-04T09:36:29.8354340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 508, in forward 2025-12-04T09:36:29.8354750Z query_vectors = self.query(hidden_states) 2025-12-04T09:36:29.8354884Z 2025-12-04T09:36:29.8354965Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8355161Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8355389Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8355882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8356342Z layer_outputs = layer_module( 2025-12-04T09:36:29.8356679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8357028Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8357434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8357827Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8358228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8358631Z self_outputs = self.self( 2025-12-04T09:36:29.8359019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8359447Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8359944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8360523Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8360761Z 2025-12-04T09:36:29.8360849Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8361056Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8361296Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8361955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8362483Z layer_outputs = layer_module( 2025-12-04T09:36:29.8362869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8363267Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8363727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8364212Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8364669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8365123Z self_outputs = self.self( 2025-12-04T09:36:29.8365550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8366041Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8366591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8367227Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8367472Z 2025-12-04T09:36:29.8367583Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8368112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8368614Z layer_outputs = layer_module( 2025-12-04T09:36:29.8368976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8369348Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8369787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8370216Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8370650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8371066Z self_outputs = self.self( 2025-12-04T09:36:29.8371485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8371943Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8372451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8373063Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8373318Z 2025-12-04T09:36:29.8373426Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8373944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8374426Z layer_outputs = layer_module( 2025-12-04T09:36:29.8374768Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8375139Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8375564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8375979Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8376397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8376808Z self_outputs = self.self( 2025-12-04T09:36:29.8377244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8377687Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8378188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8378797Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8379032Z 2025-12-04T09:36:29.8379119Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8379325Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8379560Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8380077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8380551Z layer_outputs = layer_module( 2025-12-04T09:36:29.8380900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8381265Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8381683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8382089Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8382503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8382921Z self_outputs = self.self( 2025-12-04T09:36:29.8383329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 540, in forward 2025-12-04T09:36:29.8383745Z attn_scores += diagonal_mask 2025-12-04T09:36:29.8383877Z 2025-12-04T09:36:29.8383995Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8384505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8384981Z layer_outputs = layer_module( 2025-12-04T09:36:29.8385330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8385693Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8386120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8386522Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8386934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8387338Z self_outputs = self.self( 2025-12-04T09:36:29.8387740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 578, in forward 2025-12-04T09:36:29.8388151Z attn_probs = nn.functional.softmax( 2025-12-04T09:36:29.8388290Z 2025-12-04T09:36:29.8388370Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8388585Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8388819Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8389353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8389832Z layer_outputs = layer_module( 2025-12-04T09:36:29.8390184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8390550Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8391027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8391460Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8391878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8392301Z self_outputs = self.self( 2025-12-04T09:36:29.8392709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8393218Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8393743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 862, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8394333Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-12-04T09:36:29.8394766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8395128Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8395280Z 2025-12-04T09:36:29.8395385Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8395905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8396399Z layer_outputs = layer_module( 2025-12-04T09:36:29.8396751Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8397118Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8397542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8397963Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8398382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8398795Z self_outputs = self.self( 2025-12-04T09:36:29.8399198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8399660Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8400190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 875, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8400746Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-12-04T09:36:29.8401265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 698, in _pad_and_diagonalize 2025-12-04T09:36:29.8401827Z chunked_hidden_states = nn.functional.pad( 2025-12-04T09:36:29.8402210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8402611Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8402783Z 2025-12-04T09:36:29.8402909Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8403472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8403968Z layer_outputs = layer_module( 2025-12-04T09:36:29.8404334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8404712Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8405138Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8405569Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8406034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8406451Z self_outputs = self.self( 2025-12-04T09:36:29.8406853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8407359Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8407891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8408460Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8408669Z 2025-12-04T09:36:29.8408775Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8409305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8409795Z layer_outputs = layer_module( 2025-12-04T09:36:29.8410150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8410518Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8410947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8411360Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8411764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8412167Z self_outputs = self.self( 2025-12-04T09:36:29.8412564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8413015Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8413530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8414086Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8414295Z 2025-12-04T09:36:29.8414398Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8414905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8415374Z layer_outputs = layer_module( 2025-12-04T09:36:29.8415718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8416085Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8416510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8416926Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8417336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8417818Z self_outputs = self.self( 2025-12-04T09:36:29.8418204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 617, in forward 2025-12-04T09:36:29.8418723Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-12-04T09:36:29.8418965Z 2025-12-04T09:36:29.8419046Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8419260Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8419460Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8419667Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8419916Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8420122Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8420361Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8420871Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8421398Z layer_outputs = layer_module( 2025-12-04T09:36:29.8421730Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8422103Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8422531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8422958Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8423382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8423811Z self_outputs = self.self( 2025-12-04T09:36:29.8424203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 508, in forward 2025-12-04T09:36:29.8424604Z query_vectors = self.query(hidden_states) 2025-12-04T09:36:29.8424749Z 2025-12-04T09:36:29.8424828Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8425035Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8425262Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8425751Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8426221Z layer_outputs = layer_module( 2025-12-04T09:36:29.8426568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8426929Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8427337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8427751Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8428166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8428567Z self_outputs = self.self( 2025-12-04T09:36:29.8428960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8429396Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8429891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8430753Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8431006Z 2025-12-04T09:36:29.8431085Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8431302Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8431543Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8432046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8432543Z layer_outputs = layer_module( 2025-12-04T09:36:29.8432894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8433247Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8433666Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8434085Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8434562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8434965Z self_outputs = self.self( 2025-12-04T09:36:29.8435365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8435854Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8436349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8436916Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8437169Z 2025-12-04T09:36:29.8437276Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8437803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8438304Z layer_outputs = layer_module( 2025-12-04T09:36:29.8438653Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8439029Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8439448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8439852Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8440266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8440676Z self_outputs = self.self( 2025-12-04T09:36:29.8441070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8441523Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8442130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8442766Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8443025Z 2025-12-04T09:36:29.8443137Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8443642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8444125Z layer_outputs = layer_module( 2025-12-04T09:36:29.8444475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8444845Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8445279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8445709Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8446143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8446549Z self_outputs = self.self( 2025-12-04T09:36:29.8446949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8447390Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8447884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8448452Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8448697Z 2025-12-04T09:36:29.8448838Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8449056Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8449292Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8449797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8450314Z layer_outputs = layer_module( 2025-12-04T09:36:29.8450667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8451029Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8451453Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8451881Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8452305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8452718Z self_outputs = self.self( 2025-12-04T09:36:29.8453121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 540, in forward 2025-12-04T09:36:29.8453540Z attn_scores += diagonal_mask 2025-12-04T09:36:29.8453665Z 2025-12-04T09:36:29.8453774Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8454285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8454776Z layer_outputs = layer_module( 2025-12-04T09:36:29.8455124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8455489Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8455907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8456341Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8456751Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8457155Z self_outputs = self.self( 2025-12-04T09:36:29.8457550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 578, in forward 2025-12-04T09:36:29.8457967Z attn_probs = nn.functional.softmax( 2025-12-04T09:36:29.8458098Z 2025-12-04T09:36:29.8458181Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8458384Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8458614Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8459127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8459608Z layer_outputs = layer_module( 2025-12-04T09:36:29.8459954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8460319Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8460731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8461139Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8461552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8461960Z self_outputs = self.self( 2025-12-04T09:36:29.8462353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8462845Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8463364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 862, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8463925Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-12-04T09:36:29.8464382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8464720Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8464877Z 2025-12-04T09:36:29.8464977Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8465479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8465945Z layer_outputs = layer_module( 2025-12-04T09:36:29.8466288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8466646Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8467050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8467451Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8467860Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8468264Z self_outputs = self.self( 2025-12-04T09:36:29.8468649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8469086Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8469628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 875, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8470170Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-12-04T09:36:29.8470665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 698, in _pad_and_diagonalize 2025-12-04T09:36:29.8471114Z chunked_hidden_states = nn.functional.pad( 2025-12-04T09:36:29.8471448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8471792Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8471938Z 2025-12-04T09:36:29.8472043Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8472551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8473030Z layer_outputs = layer_module( 2025-12-04T09:36:29.8473382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8473737Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8474155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8474572Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8474988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8475389Z self_outputs = self.self( 2025-12-04T09:36:29.8475788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8476241Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8476794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8477352Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8477563Z 2025-12-04T09:36:29.8477667Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8478180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8478690Z layer_outputs = layer_module( 2025-12-04T09:36:29.8479042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8479402Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8479823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8480237Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8480654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8481066Z self_outputs = self.self( 2025-12-04T09:36:29.8481477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8482035Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8482599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8483173Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8483373Z 2025-12-04T09:36:29.8483487Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8483994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8484474Z layer_outputs = layer_module( 2025-12-04T09:36:29.8484821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8485176Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8485597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8486014Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8486425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8486826Z self_outputs = self.self( 2025-12-04T09:36:29.8487217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 617, in forward 2025-12-04T09:36:29.8487745Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-12-04T09:36:29.8487985Z 2025-12-04T09:36:29.8488075Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8488288Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8488502Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8488720Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8488929Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8489131Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8489366Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8489891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8490403Z layer_outputs = layer_module( 2025-12-04T09:36:29.8490825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8491254Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8491674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8492110Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8492537Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8493009Z self_outputs = self.self( 2025-12-04T09:36:29.8493409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 508, in forward 2025-12-04T09:36:29.8493842Z query_vectors = self.query(hidden_states) 2025-12-04T09:36:29.8493983Z 2025-12-04T09:36:29.8494070Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8494274Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8494515Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8495039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8495533Z layer_outputs = layer_module( 2025-12-04T09:36:29.8495880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8496255Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8496681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8497103Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8497519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8497934Z self_outputs = self.self( 2025-12-04T09:36:29.8498342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8498784Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8499294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8499893Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8500138Z 2025-12-04T09:36:29.8500229Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8500439Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8500677Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8501172Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8501645Z layer_outputs = layer_module( 2025-12-04T09:36:29.8501975Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8502328Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8502735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8503141Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8503536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8503935Z self_outputs = self.self( 2025-12-04T09:36:29.8504322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8504748Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8505288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8505871Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8506111Z 2025-12-04T09:36:29.8506223Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8506766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8507233Z layer_outputs = layer_module( 2025-12-04T09:36:29.8507570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8507919Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8508314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8508714Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8509118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8509515Z self_outputs = self.self( 2025-12-04T09:36:29.8509893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8510318Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8510795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8511351Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8511593Z 2025-12-04T09:36:29.8511696Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8512201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8512667Z layer_outputs = layer_module( 2025-12-04T09:36:29.8512998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8513347Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8513753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8514154Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8514550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8514956Z self_outputs = self.self( 2025-12-04T09:36:29.8515361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8515788Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8516260Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8516814Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8517050Z 2025-12-04T09:36:29.8517141Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8517345Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8517577Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8518074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8518537Z layer_outputs = layer_module( 2025-12-04T09:36:29.8518908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8519263Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8519674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8520087Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8520530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8520936Z self_outputs = self.self( 2025-12-04T09:36:29.8521335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 540, in forward 2025-12-04T09:36:29.8521815Z attn_scores += diagonal_mask 2025-12-04T09:36:29.8521949Z 2025-12-04T09:36:29.8522057Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8522572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8523057Z layer_outputs = layer_module( 2025-12-04T09:36:29.8523408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8523787Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8524227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8524671Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8525078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8525492Z self_outputs = self.self( 2025-12-04T09:36:29.8525891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 578, in forward 2025-12-04T09:36:29.8526304Z attn_probs = nn.functional.softmax( 2025-12-04T09:36:29.8526448Z 2025-12-04T09:36:29.8526527Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8526742Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8526977Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8527477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8527959Z layer_outputs = layer_module( 2025-12-04T09:36:29.8528302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8528665Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8529073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8529488Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8529900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8530461Z self_outputs = self.self( 2025-12-04T09:36:29.8530867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8531339Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8531866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 862, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8532444Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-12-04T09:36:29.8532880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8533301Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8533453Z 2025-12-04T09:36:29.8533564Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8534067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8534549Z layer_outputs = layer_module( 2025-12-04T09:36:29.8534950Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8535313Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8535723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8536144Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8536559Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8536956Z self_outputs = self.self( 2025-12-04T09:36:29.8537344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8537786Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8538299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 875, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8538822Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-12-04T09:36:29.8539317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 698, in _pad_and_diagonalize 2025-12-04T09:36:29.8539768Z chunked_hidden_states = nn.functional.pad( 2025-12-04T09:36:29.8540106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8540445Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8540598Z 2025-12-04T09:36:29.8540699Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8541195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8541665Z layer_outputs = layer_module( 2025-12-04T09:36:29.8541995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8542341Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8542745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8543144Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8543553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8543949Z self_outputs = self.self( 2025-12-04T09:36:29.8544332Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8544764Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8545275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8545425Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8545437Z 2025-12-04T09:36:29.8545538Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8545874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8545989Z layer_outputs = layer_module( 2025-12-04T09:36:29.8546214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8546290Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8546580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8546700Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8546974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8547043Z self_outputs = self.self( 2025-12-04T09:36:29.8547318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8547437Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8547782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8547934Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8547937Z 2025-12-04T09:36:29.8548038Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8548386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8548462Z layer_outputs = layer_module( 2025-12-04T09:36:29.8548678Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8548762Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8549035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8549111Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8549390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8549457Z self_outputs = self.self( 2025-12-04T09:36:29.8549725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 617, in forward 2025-12-04T09:36:29.8549921Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-12-04T09:36:29.8549925Z 2025-12-04T09:36:29.8550002Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8550085Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8550161Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8550235Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8550319Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8550393Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8550498Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8550852Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8550922Z layer_outputs = layer_module( 2025-12-04T09:36:29.8551148Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8551228Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8551500Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8551582Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8551855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8551928Z self_outputs = self.self( 2025-12-04T09:36:29.8552235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 508, in forward 2025-12-04T09:36:29.8552322Z query_vectors = self.query(hidden_states) 2025-12-04T09:36:29.8552325Z 2025-12-04T09:36:29.8552407Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8552480Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8552612Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8552959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8553028Z layer_outputs = layer_module( 2025-12-04T09:36:29.8553249Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8553326Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8553597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8553681Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8553951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8554029Z self_outputs = self.self( 2025-12-04T09:36:29.8554297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8554398Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8554731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8554908Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8554912Z 2025-12-04T09:36:29.8554997Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8555074Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8555176Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8555531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8555604Z layer_outputs = layer_module( 2025-12-04T09:36:29.8555823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8555910Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8556190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8556272Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8556558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8556626Z self_outputs = self.self( 2025-12-04T09:36:29.8556902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8557002Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8557333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8557517Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8557521Z 2025-12-04T09:36:29.8557626Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8557983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8558848Z layer_outputs = layer_module( 2025-12-04T09:36:29.8559074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8559159Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8559436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8559561Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8559837Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8559908Z self_outputs = self.self( 2025-12-04T09:36:29.8560203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8560305Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8560666Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8560846Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8560850Z 2025-12-04T09:36:29.8560955Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8561325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8561398Z layer_outputs = layer_module( 2025-12-04T09:36:29.8561697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8561784Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8562066Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8562147Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8562425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8562497Z self_outputs = self.self( 2025-12-04T09:36:29.8562786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8562889Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8563247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8563429Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8563433Z 2025-12-04T09:36:29.8563513Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8563605Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8563711Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8564082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8564157Z layer_outputs = layer_module( 2025-12-04T09:36:29.8564393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8564483Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8564770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8564854Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8565141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8565248Z self_outputs = self.self( 2025-12-04T09:36:29.8565538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 540, in forward 2025-12-04T09:36:29.8565611Z attn_scores += diagonal_mask 2025-12-04T09:36:29.8565614Z 2025-12-04T09:36:29.8565719Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8566105Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8566176Z layer_outputs = layer_module( 2025-12-04T09:36:29.8566399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8566478Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8566753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8566837Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8567112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8567187Z self_outputs = self.self( 2025-12-04T09:36:29.8567461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 578, in forward 2025-12-04T09:36:29.8567543Z attn_probs = nn.functional.softmax( 2025-12-04T09:36:29.8567547Z 2025-12-04T09:36:29.8567633Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8567711Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8567813Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8568167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8568242Z layer_outputs = layer_module( 2025-12-04T09:36:29.8568468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8568546Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8568821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8568905Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8569179Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8569257Z self_outputs = self.self( 2025-12-04T09:36:29.8569532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8569650Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8570010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 862, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8570182Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-12-04T09:36:29.8570385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8570488Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8570491Z 2025-12-04T09:36:29.8570594Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8570963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8571036Z layer_outputs = layer_module( 2025-12-04T09:36:29.8571298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8571431Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8571700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8571781Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8572049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8572163Z self_outputs = self.self( 2025-12-04T09:36:29.8572448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8572563Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8572911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 875, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8573045Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-12-04T09:36:29.8573356Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 698, in _pad_and_diagonalize 2025-12-04T09:36:29.8573453Z chunked_hidden_states = nn.functional.pad( 2025-12-04T09:36:29.8573643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8573749Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8573753Z 2025-12-04T09:36:29.8573855Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8574210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8574287Z layer_outputs = layer_module( 2025-12-04T09:36:29.8574507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8574583Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8574860Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8574933Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8575210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8575283Z self_outputs = self.self( 2025-12-04T09:36:29.8575549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8575670Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8576010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8576167Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8576170Z 2025-12-04T09:36:29.8576269Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8576605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8576689Z layer_outputs = layer_module( 2025-12-04T09:36:29.8576906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8576989Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8577265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8577339Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8577713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8577796Z self_outputs = self.self( 2025-12-04T09:36:29.8578063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8578182Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8578554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8578707Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8578710Z 2025-12-04T09:36:29.8578809Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8579148Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8579229Z layer_outputs = layer_module( 2025-12-04T09:36:29.8579445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8579530Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8579799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8579875Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8580152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8580221Z self_outputs = self.self( 2025-12-04T09:36:29.8580507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 617, in forward 2025-12-04T09:36:29.8580691Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-12-04T09:36:29.8580694Z 2025-12-04T09:36:29.8580773Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8580858Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8580933Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8581006Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8581086Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8581161Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8581267Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8581609Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8581678Z layer_outputs = layer_module( 2025-12-04T09:36:29.8581896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8581973Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8582244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8582324Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8582592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8582669Z self_outputs = self.self( 2025-12-04T09:36:29.8582937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 508, in forward 2025-12-04T09:36:29.8583020Z query_vectors = self.query(hidden_states) 2025-12-04T09:36:29.8583023Z 2025-12-04T09:36:29.8583105Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8583179Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8583286Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8583659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8583729Z layer_outputs = layer_module( 2025-12-04T09:36:29.8583949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8584027Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8584325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8584403Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8584670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8584747Z self_outputs = self.self( 2025-12-04T09:36:29.8585016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8585116Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8585450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8585625Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8585631Z 2025-12-04T09:36:29.8585714Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8585789Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8585888Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8586232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8586300Z layer_outputs = layer_module( 2025-12-04T09:36:29.8586522Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8586599Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8586864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8586945Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8587214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8587283Z self_outputs = self.self( 2025-12-04T09:36:29.8587558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8587656Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8587992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8588165Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8588169Z 2025-12-04T09:36:29.8588270Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8588620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8588695Z layer_outputs = layer_module( 2025-12-04T09:36:29.8588919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8588999Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8589276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8589359Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8589674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8589752Z self_outputs = self.self( 2025-12-04T09:36:29.8590019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8590117Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8590486Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8590662Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8590666Z 2025-12-04T09:36:29.8590775Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8591121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8591194Z layer_outputs = layer_module( 2025-12-04T09:36:29.8591419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8591499Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8591785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8591870Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8592140Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8592216Z self_outputs = self.self( 2025-12-04T09:36:29.8592500Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8592600Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8592932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8593108Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8593111Z 2025-12-04T09:36:29.8593200Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8593278Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8593381Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8593729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8593801Z layer_outputs = layer_module( 2025-12-04T09:36:29.8594018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8594107Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8594380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8594462Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8594741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8594812Z self_outputs = self.self( 2025-12-04T09:36:29.8595090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 540, in forward 2025-12-04T09:36:29.8595162Z attn_scores += diagonal_mask 2025-12-04T09:36:29.8595165Z 2025-12-04T09:36:29.8595274Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8595662Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8595735Z layer_outputs = layer_module( 2025-12-04T09:36:29.8595967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8596047Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8596333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8596442Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8596716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8596791Z self_outputs = self.self( 2025-12-04T09:36:29.8597066Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 578, in forward 2025-12-04T09:36:29.8597143Z attn_probs = nn.functional.softmax( 2025-12-04T09:36:29.8597149Z 2025-12-04T09:36:29.8597232Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8597309Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8597419Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8597765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8597837Z layer_outputs = layer_module( 2025-12-04T09:36:29.8598060Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8598137Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8598418Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8598490Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8598769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8598846Z self_outputs = self.self( 2025-12-04T09:36:29.8599120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8599238Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8599594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 862, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8599766Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-12-04T09:36:29.8599966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8600065Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8600069Z 2025-12-04T09:36:29.8600173Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8600525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8600595Z layer_outputs = layer_module( 2025-12-04T09:36:29.8600822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8600903Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8601179Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8601261Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8601535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8601680Z self_outputs = self.self( 2025-12-04T09:36:29.8602032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8602162Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8602561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 875, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8602744Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-12-04T09:36:29.8603087Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 698, in _pad_and_diagonalize 2025-12-04T09:36:29.8603193Z chunked_hidden_states = nn.functional.pad( 2025-12-04T09:36:29.8603405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8603511Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8603515Z 2025-12-04T09:36:29.8603619Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8603965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8604047Z layer_outputs = layer_module( 2025-12-04T09:36:29.8604266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8604355Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8604631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8604706Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8604992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8605062Z self_outputs = self.self( 2025-12-04T09:36:29.8605348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8605463Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8605810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8605969Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8605972Z 2025-12-04T09:36:29.8606074Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8606424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8606496Z layer_outputs = layer_module( 2025-12-04T09:36:29.8606717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8606810Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8607089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8607164Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8607454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8607524Z self_outputs = self.self( 2025-12-04T09:36:29.8607812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8607929Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8608309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8608467Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8608471Z 2025-12-04T09:36:29.8608573Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8608932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8609035Z layer_outputs = layer_module( 2025-12-04T09:36:29.8609254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8609341Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8609622Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8609705Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8609988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8610059Z self_outputs = self.self( 2025-12-04T09:36:29.8610346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 617, in forward 2025-12-04T09:36:29.8610533Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-12-04T09:36:29.8610537Z 2025-12-04T09:36:29.8610624Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8610704Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8610780Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8610876Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8610950Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8611022Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8611130Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8611475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8611544Z layer_outputs = layer_module( 2025-12-04T09:36:29.8611766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8611847Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8612126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8612199Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8612479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8612557Z self_outputs = self.self( 2025-12-04T09:36:29.8612835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 508, in forward 2025-12-04T09:36:29.8612925Z query_vectors = self.query(hidden_states) 2025-12-04T09:36:29.8612929Z 2025-12-04T09:36:29.8613004Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8613078Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8613188Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8613543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8613615Z layer_outputs = layer_module( 2025-12-04T09:36:29.8613842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8613922Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8614242Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8614315Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8614593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8614672Z self_outputs = self.self( 2025-12-04T09:36:29.8614947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8615089Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8615427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8615605Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8615609Z 2025-12-04T09:36:29.8615693Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8615775Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8615879Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8616244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8616322Z layer_outputs = layer_module( 2025-12-04T09:36:29.8616548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8616630Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8616908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8616994Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8617282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8617367Z self_outputs = self.self( 2025-12-04T09:36:29.8617650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8617751Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8618089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8618269Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8618273Z 2025-12-04T09:36:29.8618383Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8618727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8618800Z layer_outputs = layer_module( 2025-12-04T09:36:29.8619025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8619104Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8619378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8619462Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8619736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8619814Z self_outputs = self.self( 2025-12-04T09:36:29.8620083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8620186Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8620576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8620756Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8620760Z 2025-12-04T09:36:29.8620871Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8621219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8621336Z layer_outputs = layer_module( 2025-12-04T09:36:29.8621563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8621642Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8621926Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8622000Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8622281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8622360Z self_outputs = self.self( 2025-12-04T09:36:29.8622648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8622750Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8623085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8623258Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8623261Z 2025-12-04T09:36:29.8623345Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8623420Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8623525Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8623876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8623945Z layer_outputs = layer_module( 2025-12-04T09:36:29.8624169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8624251Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8624531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8624611Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8624888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8624967Z self_outputs = self.self( 2025-12-04T09:36:29.8625244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 540, in forward 2025-12-04T09:36:29.8625319Z attn_scores += diagonal_mask 2025-12-04T09:36:29.8625322Z 2025-12-04T09:36:29.8625436Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8625778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8625852Z layer_outputs = layer_module( 2025-12-04T09:36:29.8626078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8626158Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8626445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8626517Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8626821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8626899Z self_outputs = self.self( 2025-12-04T09:36:29.8627168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 578, in forward 2025-12-04T09:36:29.8627288Z attn_probs = nn.functional.softmax( 2025-12-04T09:36:29.8627291Z 2025-12-04T09:36:29.8627369Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8627444Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8627556Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8627905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8627975Z layer_outputs = layer_module( 2025-12-04T09:36:29.8628205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8628284Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8628573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8628647Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8628935Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8629013Z self_outputs = self.self( 2025-12-04T09:36:29.8629293Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8629417Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8629769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 862, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8629942Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-12-04T09:36:29.8630258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8630365Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8630372Z 2025-12-04T09:36:29.8630487Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8630839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8630912Z layer_outputs = layer_module( 2025-12-04T09:36:29.8631141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8631222Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8631500Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8631586Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8631862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8631941Z self_outputs = self.self( 2025-12-04T09:36:29.8632217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8632335Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8632689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 875, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8632826Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-12-04T09:36:29.8633210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 698, in _pad_and_diagonalize 2025-12-04T09:36:29.8633302Z chunked_hidden_states = nn.functional.pad( 2025-12-04T09:36:29.8633493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8633598Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8633645Z 2025-12-04T09:36:29.8633749Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8634101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8634174Z layer_outputs = layer_module( 2025-12-04T09:36:29.8634393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8634480Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8634758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8634835Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8635119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8635195Z self_outputs = self.self( 2025-12-04T09:36:29.8635478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8635593Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8635942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8636104Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8636107Z 2025-12-04T09:36:29.8636212Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8636562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8636633Z layer_outputs = layer_module( 2025-12-04T09:36:29.8636855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8636942Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8637214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8637298Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8637568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8637641Z self_outputs = self.self( 2025-12-04T09:36:29.8637920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8638034Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8638387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8638538Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8638542Z 2025-12-04T09:36:29.8638644Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8638996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8639068Z layer_outputs = layer_module( 2025-12-04T09:36:29.8639319Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8639409Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8639689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8639816Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8640085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8640154Z self_outputs = self.self( 2025-12-04T09:36:29.8640435Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 617, in forward 2025-12-04T09:36:29.8640619Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-12-04T09:36:29.8640622Z 2025-12-04T09:36:29.8640716Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8640796Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8640872Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8640956Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8641032Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8641106Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8641221Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8641609Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8641696Z layer_outputs = layer_module( 2025-12-04T09:36:29.8641918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8642002Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8642320Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8642402Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8642713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8642797Z self_outputs = self.self( 2025-12-04T09:36:29.8643103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 508, in forward 2025-12-04T09:36:29.8643201Z query_vectors = self.query(hidden_states) 2025-12-04T09:36:29.8643205Z 2025-12-04T09:36:29.8643289Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8643372Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8643490Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8643884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8643965Z layer_outputs = layer_module( 2025-12-04T09:36:29.8644194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8644275Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8644570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8644651Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8644938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8645020Z self_outputs = self.self( 2025-12-04T09:36:29.8645307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8645421Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8645812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8646000Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8646012Z 2025-12-04T09:36:29.8646124Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8646203Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8646317Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8646684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8646754Z layer_outputs = layer_module( 2025-12-04T09:36:29.8646984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8647065Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8647357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8647432Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8647724Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8647802Z self_outputs = self.self( 2025-12-04T09:36:29.8648086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8648197Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8648548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8648728Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8648732Z 2025-12-04T09:36:29.8648842Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8649204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8649277Z layer_outputs = layer_module( 2025-12-04T09:36:29.8649512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8649588Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8649875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8649950Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8650239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8650316Z self_outputs = self.self( 2025-12-04T09:36:29.8650604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8650708Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8651048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8651230Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8651234Z 2025-12-04T09:36:29.8651344Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8651691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8651772Z layer_outputs = layer_module( 2025-12-04T09:36:29.8652023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8652104Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8652397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8652501Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8652770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8652845Z self_outputs = self.self( 2025-12-04T09:36:29.8653115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8653220Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8653548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8653721Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8653724Z 2025-12-04T09:36:29.8653809Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8653885Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8653995Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8654339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8654409Z layer_outputs = layer_module( 2025-12-04T09:36:29.8654633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8654710Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8654999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8655078Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8655355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8655429Z self_outputs = self.self( 2025-12-04T09:36:29.8655709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 540, in forward 2025-12-04T09:36:29.8655779Z attn_scores += diagonal_mask 2025-12-04T09:36:29.8655783Z 2025-12-04T09:36:29.8655892Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8656238Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8656313Z layer_outputs = layer_module( 2025-12-04T09:36:29.8656532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8656608Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8656891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8656964Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8657249Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8657316Z self_outputs = self.self( 2025-12-04T09:36:29.8657589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 578, in forward 2025-12-04T09:36:29.8657673Z attn_probs = nn.functional.softmax( 2025-12-04T09:36:29.8657676Z 2025-12-04T09:36:29.8657750Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8657861Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8657973Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8658316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8658394Z layer_outputs = layer_module( 2025-12-04T09:36:29.8658645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8658721Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8659000Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8659071Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8659348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8659418Z self_outputs = self.self( 2025-12-04T09:36:29.8659685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8659806Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8660142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 862, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8660312Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-12-04T09:36:29.8660512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8660609Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8660612Z 2025-12-04T09:36:29.8660719Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8661056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8661127Z layer_outputs = layer_module( 2025-12-04T09:36:29.8661349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8661427Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8661705Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8661776Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8662046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8662122Z self_outputs = self.self( 2025-12-04T09:36:29.8662394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8662512Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8662850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 875, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8662979Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-12-04T09:36:29.8663291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 698, in _pad_and_diagonalize 2025-12-04T09:36:29.8663378Z chunked_hidden_states = nn.functional.pad( 2025-12-04T09:36:29.8663565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8663667Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8663670Z 2025-12-04T09:36:29.8663770Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8664154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8664226Z layer_outputs = layer_module( 2025-12-04T09:36:29.8664440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8664571Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8664844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8664927Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8665203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8665273Z self_outputs = self.self( 2025-12-04T09:36:29.8665558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8665674Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8666027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8666178Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8666182Z 2025-12-04T09:36:29.8666289Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8666646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8666719Z layer_outputs = layer_module( 2025-12-04T09:36:29.8666945Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8667027Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8667303Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8667385Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8667665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8667737Z self_outputs = self.self( 2025-12-04T09:36:29.8668019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8668134Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8668485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8668636Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8668643Z 2025-12-04T09:36:29.8668744Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8669101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8669176Z layer_outputs = layer_module( 2025-12-04T09:36:29.8669401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8669482Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8669760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8669846Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8670157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8670236Z self_outputs = self.self( 2025-12-04T09:36:29.8670511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 617, in forward 2025-12-04T09:36:29.8670698Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-12-04T09:36:29.8670733Z 2025-12-04T09:36:29.8670823Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8670904Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8670981Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8671066Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8671140Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8671222Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8671324Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8671673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8671753Z layer_outputs = layer_module( 2025-12-04T09:36:29.8671972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8672051Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8672347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8672421Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8672706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8672775Z self_outputs = self.self( 2025-12-04T09:36:29.8673049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 508, in forward 2025-12-04T09:36:29.8673145Z query_vectors = self.query(hidden_states) 2025-12-04T09:36:29.8673149Z 2025-12-04T09:36:29.8673225Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8673308Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8673412Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8673766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8673846Z layer_outputs = layer_module( 2025-12-04T09:36:29.8674065Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8674143Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8674428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8674502Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8674800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8674870Z self_outputs = self.self( 2025-12-04T09:36:29.8675146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8675259Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8675600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8675791Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8675794Z 2025-12-04T09:36:29.8675872Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8675948Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8676097Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8676446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8676518Z layer_outputs = layer_module( 2025-12-04T09:36:29.8676756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8676872Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8677153Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8677228Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8677504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8677582Z self_outputs = self.self( 2025-12-04T09:36:29.8677859Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8677968Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8678305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8678498Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8678501Z 2025-12-04T09:36:29.8678606Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8678940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8679016Z layer_outputs = layer_module( 2025-12-04T09:36:29.8679225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8679303Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8679580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8679652Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8679920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8679996Z self_outputs = self.self( 2025-12-04T09:36:29.8680262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8680365Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8680691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8680864Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8680874Z 2025-12-04T09:36:29.8680972Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8681307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8681386Z layer_outputs = layer_module( 2025-12-04T09:36:29.8681657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8681746Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8682032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8682106Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8682506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8682587Z self_outputs = self.self( 2025-12-04T09:36:29.8682894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8683026Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8683431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8683634Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8683638Z 2025-12-04T09:36:29.8683724Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8683819Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8683935Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8684292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8684366Z layer_outputs = layer_module( 2025-12-04T09:36:29.8684598Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8684678Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8684972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8685049Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8685333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8685414Z self_outputs = self.self( 2025-12-04T09:36:29.8685700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 540, in forward 2025-12-04T09:36:29.8685782Z attn_scores += diagonal_mask 2025-12-04T09:36:29.8685785Z 2025-12-04T09:36:29.8685891Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8686246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8686331Z layer_outputs = layer_module( 2025-12-04T09:36:29.8686557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8686638Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8686934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8687010Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8687304Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8687375Z self_outputs = self.self( 2025-12-04T09:36:29.8687657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 578, in forward 2025-12-04T09:36:29.8687743Z attn_probs = nn.functional.softmax( 2025-12-04T09:36:29.8687746Z 2025-12-04T09:36:29.8687829Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8687914Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8688020Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8688376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8688456Z layer_outputs = layer_module( 2025-12-04T09:36:29.8688679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8688804Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8689102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8689178Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8689471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8689577Z self_outputs = self.self( 2025-12-04T09:36:29.8689858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8689990Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8690347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 862, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8690534Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-12-04T09:36:29.8690741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8690846Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8690850Z 2025-12-04T09:36:29.8690967Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8691347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8691439Z layer_outputs = layer_module( 2025-12-04T09:36:29.8691652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8691730Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8692007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8692080Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8692346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8692422Z self_outputs = self.self( 2025-12-04T09:36:29.8692689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8692813Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8693151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 875, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8693282Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-12-04T09:36:29.8693599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 698, in _pad_and_diagonalize 2025-12-04T09:36:29.8693689Z chunked_hidden_states = nn.functional.pad( 2025-12-04T09:36:29.8693884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8693979Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8693983Z 2025-12-04T09:36:29.8694086Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8694430Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8694500Z layer_outputs = layer_module( 2025-12-04T09:36:29.8694720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8694797Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8695100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8695182Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8695449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8695517Z self_outputs = self.self( 2025-12-04T09:36:29.8695821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8695932Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8696280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8696427Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8696431Z 2025-12-04T09:36:29.8696533Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8696878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8696947Z layer_outputs = layer_module( 2025-12-04T09:36:29.8697163Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8697242Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8697510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8697590Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8697856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8697931Z self_outputs = self.self( 2025-12-04T09:36:29.8698198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8698308Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8698653Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8698800Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8698804Z 2025-12-04T09:36:29.8698912Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8699247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8699317Z layer_outputs = layer_module( 2025-12-04T09:36:29.8699538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8699615Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8699886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8699966Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8700235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8700313Z self_outputs = self.self( 2025-12-04T09:36:29.8700589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 617, in forward 2025-12-04T09:36:29.8700768Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-12-04T09:36:29.8700772Z 2025-12-04T09:36:29.8700857Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8700934Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8701055Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8701132Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8701205Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8701285Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8701386Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8701723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8701839Z layer_outputs = layer_module( 2025-12-04T09:36:29.8702056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8702145Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8702413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8702490Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8702767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8702834Z self_outputs = self.self( 2025-12-04T09:36:29.8703104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 508, in forward 2025-12-04T09:36:29.8703199Z query_vectors = self.query(hidden_states) 2025-12-04T09:36:29.8703202Z 2025-12-04T09:36:29.8703278Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8703362Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8703461Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8703796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8703874Z layer_outputs = layer_module( 2025-12-04T09:36:29.8704088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8704172Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8704443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8704518Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8704793Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8704860Z self_outputs = self.self( 2025-12-04T09:36:29.8705128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8705236Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8705564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8705747Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8705751Z 2025-12-04T09:36:29.8705825Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8705899Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8706010Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8706345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8706422Z layer_outputs = layer_module( 2025-12-04T09:36:29.8706634Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8706710Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8707026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8707103Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8707386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8707455Z self_outputs = self.self( 2025-12-04T09:36:29.8707762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8707873Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8708214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8708391Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8708401Z 2025-12-04T09:36:29.8708505Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8708837Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8708913Z layer_outputs = layer_module( 2025-12-04T09:36:29.8709124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8709203Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8709476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8709547Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8709819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8709888Z self_outputs = self.self( 2025-12-04T09:36:29.8710157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8710263Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8710587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8710771Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8710776Z 2025-12-04T09:36:29.8710875Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8711211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8711288Z layer_outputs = layer_module( 2025-12-04T09:36:29.8711501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8711585Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8711849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8711922Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8712198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8712265Z self_outputs = self.self( 2025-12-04T09:36:29.8712531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 523, in forward 2025-12-04T09:36:29.8712635Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-12-04T09:36:29.8712998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 795, in _sliding_chunks_query_key_matmul 2025-12-04T09:36:29.8713178Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-12-04T09:36:29.8713182Z 2025-12-04T09:36:29.8713259Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8713334Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8713438Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8713811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8713888Z layer_outputs = layer_module( 2025-12-04T09:36:29.8714102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8714178Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8714456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8714526Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8714801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8714877Z self_outputs = self.self( 2025-12-04T09:36:29.8715151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 540, in forward 2025-12-04T09:36:29.8715235Z attn_scores += diagonal_mask 2025-12-04T09:36:29.8715239Z 2025-12-04T09:36:29.8715342Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8715690Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8715771Z layer_outputs = layer_module( 2025-12-04T09:36:29.8715995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8716081Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8716363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8716437Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8716721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8716792Z self_outputs = self.self( 2025-12-04T09:36:29.8717078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 578, in forward 2025-12-04T09:36:29.8717158Z attn_probs = nn.functional.softmax( 2025-12-04T09:36:29.8717161Z 2025-12-04T09:36:29.8717240Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8717325Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8717430Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8717777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8717857Z layer_outputs = layer_module( 2025-12-04T09:36:29.8718076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8718165Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8718440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8718513Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8718797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8718867Z self_outputs = self.self( 2025-12-04T09:36:29.8719173Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8719302Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8719655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 862, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8719865Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-12-04T09:36:29.8720059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8720165Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8720177Z 2025-12-04T09:36:29.8720276Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8720615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8720694Z layer_outputs = layer_module( 2025-12-04T09:36:29.8720913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8720992Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8721279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8721354Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8721816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8721898Z self_outputs = self.self( 2025-12-04T09:36:29.8722183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8722318Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8722675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 875, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8722823Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-12-04T09:36:29.8723164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 698, in _pad_and_diagonalize 2025-12-04T09:36:29.8723258Z chunked_hidden_states = nn.functional.pad( 2025-12-04T09:36:29.8723460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 5418, in pad 2025-12-04T09:36:29.8723555Z return torch._C._nn.pad(input, pad, mode, value) 2025-12-04T09:36:29.8723559Z 2025-12-04T09:36:29.8723659Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8724024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8724095Z layer_outputs = layer_module( 2025-12-04T09:36:29.8724314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8724391Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8724677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8724763Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8725047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8725126Z self_outputs = self.self( 2025-12-04T09:36:29.8725406Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8725559Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8725923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8726077Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8726114Z 2025-12-04T09:36:29.8726228Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8726582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8726656Z layer_outputs = layer_module( 2025-12-04T09:36:29.8726888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8726970Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8727265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8727342Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8727631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8727714Z self_outputs = self.self( 2025-12-04T09:36:29.8727997Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 612, in forward 2025-12-04T09:36:29.8728112Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-12-04T09:36:29.8728473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 877, in _sliding_chunks_matmul_attn_probs_value 2025-12-04T09:36:29.8728624Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-12-04T09:36:29.8728628Z 2025-12-04T09:36:29.8728740Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:29.8729091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1258, in torch_dynamo_resume_in_forward_at_1243 2025-12-04T09:36:29.8729163Z layer_outputs = layer_module( 2025-12-04T09:36:29.8729394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:36:29.8729475Z return super().__call__(*args, **kwargs) 2025-12-04T09:36:29.8729764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1198, in forward 2025-12-04T09:36:29.8729841Z self_attn_outputs = self.attention( 2025-12-04T09:36:29.8730266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1134, in forward 2025-12-04T09:36:29.8730359Z self_outputs = self.self( 2025-12-04T09:36:29.8730654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 617, in forward 2025-12-04T09:36:29.8730860Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-12-04T09:36:29.8730864Z 2025-12-04T09:36:29.8730948Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8731034Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8731123Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8731206Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8731286Z cudagraph partition due to non gpu ops 2025-12-04T09:36:29.8731378Z cudagraph partition due to non gpu ops 2025-12-04T09:36:45.0770119Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:45.0774569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1715, in torch_dynamo_resume_in_forward_at_1702 2025-12-04T09:36:45.0776275Z prediction_scores = self.lm_head(sequence_output) 2025-12-04T09:36:45.0776839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1332, in forward 2025-12-04T09:36:45.0777247Z x = self.dense(features) 2025-12-04T09:36:45.0777446Z 2025-12-04T09:36:45.0781503Z cudagraph partition due to non gpu ops 2025-12-04T09:36:45.0787555Z cudagraph partition due to non gpu ops 2025-12-04T09:36:45.0789193Z cudagraph partition due to non gpu ops 2025-12-04T09:36:45.0789404Z cudagraph partition due to non gpu ops 2025-12-04T09:36:45.0789648Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:36:45.0790180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/longformer/modeling_longformer.py", line 1722, in torch_dynamo_resume_in_forward_at_1702 2025-12-04T09:36:45.0790764Z masked_lm_loss = loss_fct(prediction_scores.view(-1, self.config.vocab_size), labels.view(-1)) 2025-12-04T09:36:45.0791032Z 2025-12-04T09:36:46.4579349Z Compilation time (from dynamo_timed): 48.956717397 2025-12-04T09:36:46.4718837Z pass 2025-12-04T09:36:46.4719479Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:36:46.4720447Z TIMING: _recursive_pre_grad_passes:0.10869 _recursive_joint_graph_passes:0.92799 inductor_compile:30.42739 backend_compile:43.47218 gc:0.00419 entire_frame_compile:48.95672 _recursive_post_grad_passes:0.93538 async_compile.wait:4.79814 code_gen:25.9746 total_wall_time:48.95672 2025-12-04T09:36:46.4721422Z STATS: call_* op count: 1789 | FakeTensorMode.__torch_dispatch__:58350 | FakeTensor.__torch_dispatch__:8508 | ProxyTorchDispatchMode.__torch_dispatch__:12191 2025-12-04T09:36:46.4722003Z Dynamo produced 5 graphs covering 1789 ops with 4 graph breaks (1 unique) 2025-12-04T09:36:50.1215320Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:36:50.1216168Z import pynvml # type: ignore[import] 2025-12-04T09:36:53.3988099Z 2025-12-04T09:36:55.5909758Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:36:55.5910178Z loading model: 0it [00:02, ?it/s] 2025-12-04T09:36:55.5910533Z cpu eval BartForCausalLM 2025-12-04T09:36:57.3132636Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:36:57.6436537Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:36:57.9796883Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:37:07.3991244Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.3994456Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4014714Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4015077Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4015350Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4015572Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4015786Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4015995Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4016225Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4016439Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4016646Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4016873Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4017101Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4017338Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4017545Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4017760Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4017971Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4019496Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4019743Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4020020Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:07.4020437Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:07.4020812Z res = mod(**inputs) 2025-12-04T09:37:07.4021248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1902, in forward 2025-12-04T09:37:07.4021785Z outputs = self.model.decoder( 2025-12-04T09:37:07.4022202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1122, in forward 2025-12-04T09:37:07.4022645Z layer_outputs = decoder_layer( 2025-12-04T09:37:07.4023014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:07.4023390Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:07.4023803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4024193Z return func(*args, **kwargs) 2025-12-04T09:37:07.4024584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 418, in forward 2025-12-04T09:37:07.4025004Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:37:07.4025424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4025807Z return func(*args, **kwargs) 2025-12-04T09:37:07.4026185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 258, in forward 2025-12-04T09:37:07.4026594Z attn_output, attn_weights = attention_interface( 2025-12-04T09:37:07.4027047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:37:07.4027531Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:07.4027720Z 2025-12-04T09:37:07.4027803Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4028017Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4028226Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4028436Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4028633Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4028836Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4029035Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4029231Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4029434Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4029641Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4029839Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4030055Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4030558Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4030813Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:07.4031214Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:07.4031573Z res = mod(**inputs) 2025-12-04T09:37:07.4031977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1902, in forward 2025-12-04T09:37:07.4032398Z outputs = self.model.decoder( 2025-12-04T09:37:07.4032809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1122, in forward 2025-12-04T09:37:07.4033206Z layer_outputs = decoder_layer( 2025-12-04T09:37:07.4033569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:07.4033955Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:07.4034420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4034821Z return func(*args, **kwargs) 2025-12-04T09:37:07.4035191Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 418, in forward 2025-12-04T09:37:07.4035602Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:37:07.4036063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4036444Z return func(*args, **kwargs) 2025-12-04T09:37:07.4036809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 258, in forward 2025-12-04T09:37:07.4037249Z attn_output, attn_weights = attention_interface( 2025-12-04T09:37:07.4037698Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:37:07.4038174Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:07.4038365Z 2025-12-04T09:37:07.4038448Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4038663Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4038872Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4039073Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4039286Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4039490Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4039688Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4039895Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4040102Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4040299Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4040508Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4040716Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4040913Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4041160Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:07.4041523Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:07.4041938Z res = mod(**inputs) 2025-12-04T09:37:07.4042314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1902, in forward 2025-12-04T09:37:07.4042748Z outputs = self.model.decoder( 2025-12-04T09:37:07.4043152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1122, in forward 2025-12-04T09:37:07.4043590Z layer_outputs = decoder_layer( 2025-12-04T09:37:07.4043976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:07.4044391Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:07.4044794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4045205Z return func(*args, **kwargs) 2025-12-04T09:37:07.4045605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 418, in forward 2025-12-04T09:37:07.4046040Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:37:07.4046463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4046865Z return func(*args, **kwargs) 2025-12-04T09:37:07.4047262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 258, in forward 2025-12-04T09:37:07.4047690Z attn_output, attn_weights = attention_interface( 2025-12-04T09:37:07.4048172Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:37:07.4048693Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:07.4048962Z 2025-12-04T09:37:07.4049058Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4049278Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4049505Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4049726Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4049939Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4050202Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4050428Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4050645Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4050867Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4051091Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4051311Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4051527Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4051813Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4052079Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:07.4052444Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:07.4052780Z res = mod(**inputs) 2025-12-04T09:37:07.4053152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1902, in forward 2025-12-04T09:37:07.4053549Z outputs = self.model.decoder( 2025-12-04T09:37:07.4053935Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1122, in forward 2025-12-04T09:37:07.4054329Z layer_outputs = decoder_layer( 2025-12-04T09:37:07.4054692Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:07.4055065Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:07.4055457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4055837Z return func(*args, **kwargs) 2025-12-04T09:37:07.4056221Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 418, in forward 2025-12-04T09:37:07.4056638Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:37:07.4057036Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4057409Z return func(*args, **kwargs) 2025-12-04T09:37:07.4057766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 258, in forward 2025-12-04T09:37:07.4058170Z attn_output, attn_weights = attention_interface( 2025-12-04T09:37:07.4058613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:37:07.4059091Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:07.4059275Z 2025-12-04T09:37:07.4059358Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4059577Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4059790Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4059999Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4060203Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4060411Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4060634Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4060831Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4061037Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4061243Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4061440Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4061646Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4061852Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4062080Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:07.4062481Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:07.4062807Z res = mod(**inputs) 2025-12-04T09:37:07.4063166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1902, in forward 2025-12-04T09:37:07.4063542Z outputs = self.model.decoder( 2025-12-04T09:37:07.4063931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1122, in forward 2025-12-04T09:37:07.4064365Z layer_outputs = decoder_layer( 2025-12-04T09:37:07.4064717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:07.4065088Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:07.4065475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4065852Z return func(*args, **kwargs) 2025-12-04T09:37:07.4066221Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 418, in forward 2025-12-04T09:37:07.4066639Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:37:07.4067043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4067416Z return func(*args, **kwargs) 2025-12-04T09:37:07.4067787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 258, in forward 2025-12-04T09:37:07.4068199Z attn_output, attn_weights = attention_interface( 2025-12-04T09:37:07.4068646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:37:07.4069122Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:07.4069312Z 2025-12-04T09:37:07.4069394Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4069614Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4069837Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4070052Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4070274Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4070495Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4070708Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4070929Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4071150Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4071364Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4071587Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4071805Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4072017Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4072272Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:07.4072665Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:07.4073014Z res = mod(**inputs) 2025-12-04T09:37:07.4073398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1902, in forward 2025-12-04T09:37:07.4073813Z outputs = self.model.decoder( 2025-12-04T09:37:07.4074223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1122, in forward 2025-12-04T09:37:07.4074639Z layer_outputs = decoder_layer( 2025-12-04T09:37:07.4075011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:07.4075411Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:07.4075826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4076228Z return func(*args, **kwargs) 2025-12-04T09:37:07.4076668Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 418, in forward 2025-12-04T09:37:07.4077114Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:37:07.4077544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4077934Z return func(*args, **kwargs) 2025-12-04T09:37:07.4078329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 258, in forward 2025-12-04T09:37:07.4078807Z attn_output, attn_weights = attention_interface( 2025-12-04T09:37:07.4079279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:37:07.4079794Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:07.4079995Z 2025-12-04T09:37:07.4080082Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4080311Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4080531Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4080752Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4080971Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4081181Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4081400Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4081701Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4081929Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4082153Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4082374Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4082597Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4082811Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4083067Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:07.4083460Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:07.4083801Z res = mod(**inputs) 2025-12-04T09:37:07.4084195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1902, in forward 2025-12-04T09:37:07.4084617Z outputs = self.model.decoder( 2025-12-04T09:37:07.4085027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1122, in forward 2025-12-04T09:37:07.4085432Z layer_outputs = decoder_layer( 2025-12-04T09:37:07.4085816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:07.4086211Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:07.4086612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4087025Z return func(*args, **kwargs) 2025-12-04T09:37:07.4087423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 418, in forward 2025-12-04T09:37:07.4087862Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:37:07.4088282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4088678Z return func(*args, **kwargs) 2025-12-04T09:37:07.4089076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 258, in forward 2025-12-04T09:37:07.4089510Z attn_output, attn_weights = attention_interface( 2025-12-04T09:37:07.4089991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:37:07.4090499Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:07.4090691Z 2025-12-04T09:37:07.4090783Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4091003Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4091227Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4091491Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4091709Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4091931Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4092150Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4092371Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4092585Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4092864Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4093087Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4093304Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4093526Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4093778Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:07.4094162Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:07.4094509Z res = mod(**inputs) 2025-12-04T09:37:07.4094904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1902, in forward 2025-12-04T09:37:07.4095321Z outputs = self.model.decoder( 2025-12-04T09:37:07.4095727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1122, in forward 2025-12-04T09:37:07.4096150Z layer_outputs = decoder_layer( 2025-12-04T09:37:07.4096532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:07.4096895Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:07.4097267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4097636Z return func(*args, **kwargs) 2025-12-04T09:37:07.4098001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 418, in forward 2025-12-04T09:37:07.4098408Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:37:07.4098799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4099167Z return func(*args, **kwargs) 2025-12-04T09:37:07.4099534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 258, in forward 2025-12-04T09:37:07.4099929Z attn_output, attn_weights = attention_interface( 2025-12-04T09:37:07.4100371Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:37:07.4100844Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:07.4101023Z 2025-12-04T09:37:07.4101113Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4101315Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4101525Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4101731Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4101931Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4102136Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4102343Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4102547Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4102745Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4102949Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4103157Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4103357Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4103558Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4103793Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:07.4104144Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:07.4104465Z res = mod(**inputs) 2025-12-04T09:37:07.4104822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1902, in forward 2025-12-04T09:37:07.4105280Z outputs = self.model.decoder( 2025-12-04T09:37:07.4105652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1122, in forward 2025-12-04T09:37:07.4106035Z layer_outputs = decoder_layer( 2025-12-04T09:37:07.4106386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:07.4106784Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:07.4107166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4107539Z return func(*args, **kwargs) 2025-12-04T09:37:07.4107910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 418, in forward 2025-12-04T09:37:07.4108309Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:37:07.4108711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4109086Z return func(*args, **kwargs) 2025-12-04T09:37:07.4109444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 258, in forward 2025-12-04T09:37:07.4109848Z attn_output, attn_weights = attention_interface( 2025-12-04T09:37:07.4110292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:37:07.4110769Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:07.4110947Z 2025-12-04T09:37:07.4111027Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4111242Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4111450Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4111654Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4111851Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4112059Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4112262Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4112460Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4112663Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4112869Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4113069Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4113278Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4113484Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4113712Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:07.4114076Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:07.4114400Z res = mod(**inputs) 2025-12-04T09:37:07.4114761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1902, in forward 2025-12-04T09:37:07.4115144Z outputs = self.model.decoder( 2025-12-04T09:37:07.4115532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1122, in forward 2025-12-04T09:37:07.4115931Z layer_outputs = decoder_layer( 2025-12-04T09:37:07.4116296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:07.4116664Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:07.4117059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4117443Z return func(*args, **kwargs) 2025-12-04T09:37:07.4117815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 418, in forward 2025-12-04T09:37:07.4118231Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:37:07.4118642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4119058Z return func(*args, **kwargs) 2025-12-04T09:37:07.4119429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 258, in forward 2025-12-04T09:37:07.4119844Z attn_output, attn_weights = attention_interface( 2025-12-04T09:37:07.4120299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:37:07.4120843Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:07.4121053Z 2025-12-04T09:37:07.4121141Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4121368Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4121594Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4121906Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4122140Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4122372Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4122599Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4122831Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4123068Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4123272Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4123484Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4123698Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4123904Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4124149Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:07.4124521Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:07.4124856Z res = mod(**inputs) 2025-12-04T09:37:07.4125218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1902, in forward 2025-12-04T09:37:07.4125611Z outputs = self.model.decoder( 2025-12-04T09:37:07.4126002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1122, in forward 2025-12-04T09:37:07.4126386Z layer_outputs = decoder_layer( 2025-12-04T09:37:07.4126747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:07.4127143Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:07.4127526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4127909Z return func(*args, **kwargs) 2025-12-04T09:37:07.4128290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 418, in forward 2025-12-04T09:37:07.4128708Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:37:07.4129110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4129492Z return func(*args, **kwargs) 2025-12-04T09:37:07.4129873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 258, in forward 2025-12-04T09:37:07.4130475Z attn_output, attn_weights = attention_interface( 2025-12-04T09:37:07.4130938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:37:07.4131445Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:07.4131622Z 2025-12-04T09:37:07.4131709Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4131908Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4132117Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4132322Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4132526Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4132721Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4132922Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4133219Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4133417Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4133623Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4133820Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4134014Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4134215Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4134489Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:07.4134832Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:07.4135143Z res = mod(**inputs) 2025-12-04T09:37:07.4135492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1902, in forward 2025-12-04T09:37:07.4135866Z outputs = self.model.decoder( 2025-12-04T09:37:07.4136227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1122, in forward 2025-12-04T09:37:07.4136598Z layer_outputs = decoder_layer( 2025-12-04T09:37:07.4136942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:07.4137288Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:07.4137661Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4138022Z return func(*args, **kwargs) 2025-12-04T09:37:07.4138378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 418, in forward 2025-12-04T09:37:07.4138767Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:37:07.4139152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:07.4139509Z return func(*args, **kwargs) 2025-12-04T09:37:07.4139867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 258, in forward 2025-12-04T09:37:07.4140251Z attn_output, attn_weights = attention_interface( 2025-12-04T09:37:07.4140682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:37:07.4141138Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:07.4141311Z 2025-12-04T09:37:07.4141387Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4141593Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4141794Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4141993Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4142186Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4142383Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4142581Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4142772Z cudagraph partition due to non gpu ops 2025-12-04T09:37:07.4143003Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:07.4143354Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:07.4143656Z res = mod(**inputs) 2025-12-04T09:37:07.4144003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bart/modeling_bart.py", line 1924, in forward 2025-12-04T09:37:07.4144440Z loss = loss_fct(logits.view(-1, self.config.vocab_size), labels.view(-1)) 2025-12-04T09:37:07.4144627Z 2025-12-04T09:37:17.4135764Z Compilation time (from dynamo_timed): 18.459368577 2025-12-04T09:37:17.4385158Z pass 2025-12-04T09:37:17.4385859Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:37:17.4387127Z TIMING: _recursive_pre_grad_passes:0.03015 _recursive_joint_graph_passes:0.6769 _recursive_post_grad_passes:0.05987 async_compile.wait:1.10476 code_gen:9.68041 inductor_compile:11.28824 backend_compile:16.00606 gc:0.00026 entire_frame_compile:18.45937 total_wall_time:18.45937 2025-12-04T09:37:17.4388165Z STATS: call_* op count: 338 | FakeTensorMode.__torch_dispatch__:20382 | FakeTensor.__torch_dispatch__:2899 | ProxyTorchDispatchMode.__torch_dispatch__:4089 2025-12-04T09:37:17.4388729Z Dynamo produced 1 graphs covering 338 ops with 0 graph breaks (0 unique) 2025-12-04T09:37:19.9606848Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:37:19.9608619Z import pynvml # type: ignore[import] 2025-12-04T09:37:23.1326001Z 2025-12-04T09:37:24.1435533Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:37:24.1436091Z loading model: 0it [00:01, ?it/s] 2025-12-04T09:37:24.1436994Z cpu eval BertForMaskedLM 2025-12-04T09:37:24.6443810Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:37:24.7635413Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:37:24.8912529Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:37:34.3010869Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3011327Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3011664Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3012458Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3012884Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3013282Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3013647Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3013984Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3014536Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3014852Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3015085Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3015350Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3015569Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3015770Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3015983Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3016187Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3016397Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3016613Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3016909Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3017144Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:34.3017538Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:34.3017893Z res = mod(**inputs) 2025-12-04T09:37:34.3018294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1280, in forward 2025-12-04T09:37:34.3018689Z outputs = self.bert( 2025-12-04T09:37:34.3019064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1000, in forward 2025-12-04T09:37:34.3019458Z encoder_outputs = self.encoder( 2025-12-04T09:37:34.3019839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 650, in forward 2025-12-04T09:37:34.3020230Z layer_outputs = layer_module( 2025-12-04T09:37:34.3020584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:34.3020963Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:34.3021354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3021742Z return func(*args, **kwargs) 2025-12-04T09:37:34.3022457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 558, in forward 2025-12-04T09:37:34.3022883Z self_attention_outputs = self.attention( 2025-12-04T09:37:34.3023261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3023633Z return func(*args, **kwargs) 2025-12-04T09:37:34.3024003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 488, in forward 2025-12-04T09:37:34.3024475Z self_outputs = self.self( 2025-12-04T09:37:34.3024841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3025221Z return func(*args, **kwargs) 2025-12-04T09:37:34.3025597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 413, in forward 2025-12-04T09:37:34.3026105Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:34.3026309Z 2025-12-04T09:37:34.3026394Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3026610Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3026812Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3027019Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3027224Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3027434Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3027630Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3027833Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3028035Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3028229Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3028432Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3028637Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3028834Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3029069Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:34.3029444Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:34.3029811Z res = mod(**inputs) 2025-12-04T09:37:34.3030372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1280, in forward 2025-12-04T09:37:34.3030774Z outputs = self.bert( 2025-12-04T09:37:34.3031145Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1000, in forward 2025-12-04T09:37:34.3031532Z encoder_outputs = self.encoder( 2025-12-04T09:37:34.3031916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 650, in forward 2025-12-04T09:37:34.3032295Z layer_outputs = layer_module( 2025-12-04T09:37:34.3032646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:34.3033006Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:34.3033395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3033771Z return func(*args, **kwargs) 2025-12-04T09:37:34.3034136Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 558, in forward 2025-12-04T09:37:34.3034529Z self_attention_outputs = self.attention( 2025-12-04T09:37:34.3034909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3035276Z return func(*args, **kwargs) 2025-12-04T09:37:34.3035635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 488, in forward 2025-12-04T09:37:34.3036038Z self_outputs = self.self( 2025-12-04T09:37:34.3036394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3036817Z return func(*args, **kwargs) 2025-12-04T09:37:34.3037187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 413, in forward 2025-12-04T09:37:34.3037627Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:34.3037816Z 2025-12-04T09:37:34.3037905Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3038162Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3038429Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3038635Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3038843Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3039051Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3039253Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3039460Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3039666Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3039866Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3040077Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3040286Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3040495Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3040732Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:34.3041100Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:34.3041474Z res = mod(**inputs) 2025-12-04T09:37:34.3042032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1280, in forward 2025-12-04T09:37:34.3042448Z outputs = self.bert( 2025-12-04T09:37:34.3042843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1000, in forward 2025-12-04T09:37:34.3043254Z encoder_outputs = self.encoder( 2025-12-04T09:37:34.3043650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 650, in forward 2025-12-04T09:37:34.3044038Z layer_outputs = layer_module( 2025-12-04T09:37:34.3044395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:34.3044747Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:34.3045122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3045489Z return func(*args, **kwargs) 2025-12-04T09:37:34.3045846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 558, in forward 2025-12-04T09:37:34.3046226Z self_attention_outputs = self.attention( 2025-12-04T09:37:34.3046597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3046955Z return func(*args, **kwargs) 2025-12-04T09:37:34.3047307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 488, in forward 2025-12-04T09:37:34.3047674Z self_outputs = self.self( 2025-12-04T09:37:34.3048025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3048386Z return func(*args, **kwargs) 2025-12-04T09:37:34.3048735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 413, in forward 2025-12-04T09:37:34.3049158Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:34.3049335Z 2025-12-04T09:37:34.3049420Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3049618Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3049823Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3050024Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3050215Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3050457Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3050658Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3050857Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3051047Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3051247Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3051449Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3051705Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3051905Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3052135Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:34.3052479Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:34.3052795Z res = mod(**inputs) 2025-12-04T09:37:34.3053147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1280, in forward 2025-12-04T09:37:34.3053514Z outputs = self.bert( 2025-12-04T09:37:34.3053872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1000, in forward 2025-12-04T09:37:34.3054240Z encoder_outputs = self.encoder( 2025-12-04T09:37:34.3054598Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 650, in forward 2025-12-04T09:37:34.3054953Z layer_outputs = layer_module( 2025-12-04T09:37:34.3055288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:34.3055632Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:34.3055986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3056329Z return func(*args, **kwargs) 2025-12-04T09:37:34.3056676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 558, in forward 2025-12-04T09:37:34.3057049Z self_attention_outputs = self.attention( 2025-12-04T09:37:34.3057408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3057753Z return func(*args, **kwargs) 2025-12-04T09:37:34.3058100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 488, in forward 2025-12-04T09:37:34.3058468Z self_outputs = self.self( 2025-12-04T09:37:34.3058814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3059176Z return func(*args, **kwargs) 2025-12-04T09:37:34.3059539Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 413, in forward 2025-12-04T09:37:34.3059953Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:34.3060127Z 2025-12-04T09:37:34.3060205Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3060452Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3060668Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3060862Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3061062Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3061262Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3061456Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3061657Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3061857Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3062057Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3062247Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3062444Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3062646Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3062871Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:34.3063224Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:34.3063578Z res = mod(**inputs) 2025-12-04T09:37:34.3063922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1280, in forward 2025-12-04T09:37:34.3064288Z outputs = self.bert( 2025-12-04T09:37:34.3064627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1000, in forward 2025-12-04T09:37:34.3065049Z encoder_outputs = self.encoder( 2025-12-04T09:37:34.3065399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 650, in forward 2025-12-04T09:37:34.3065759Z layer_outputs = layer_module( 2025-12-04T09:37:34.3066088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:34.3066431Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:34.3066803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3067161Z return func(*args, **kwargs) 2025-12-04T09:37:34.3067526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 558, in forward 2025-12-04T09:37:34.3067890Z self_attention_outputs = self.attention( 2025-12-04T09:37:34.3068254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3068606Z return func(*args, **kwargs) 2025-12-04T09:37:34.3068956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 488, in forward 2025-12-04T09:37:34.3069307Z self_outputs = self.self( 2025-12-04T09:37:34.3069650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3070002Z return func(*args, **kwargs) 2025-12-04T09:37:34.3070345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 413, in forward 2025-12-04T09:37:34.3070755Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:34.3070930Z 2025-12-04T09:37:34.3071006Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3071204Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3071398Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3071593Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3071788Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3071975Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3072170Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3072366Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3072553Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3072747Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3072940Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3073140Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3073328Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3073550Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:34.3073891Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:34.3074190Z res = mod(**inputs) 2025-12-04T09:37:34.3074530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1280, in forward 2025-12-04T09:37:34.3074887Z outputs = self.bert( 2025-12-04T09:37:34.3075222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1000, in forward 2025-12-04T09:37:34.3075585Z encoder_outputs = self.encoder( 2025-12-04T09:37:34.3075936Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 650, in forward 2025-12-04T09:37:34.3076294Z layer_outputs = layer_module( 2025-12-04T09:37:34.3076647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:34.3076993Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:34.3077358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3077718Z return func(*args, **kwargs) 2025-12-04T09:37:34.3078101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 558, in forward 2025-12-04T09:37:34.3078481Z self_attention_outputs = self.attention( 2025-12-04T09:37:34.3078855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3079209Z return func(*args, **kwargs) 2025-12-04T09:37:34.3079567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 488, in forward 2025-12-04T09:37:34.3079939Z self_outputs = self.self( 2025-12-04T09:37:34.3080291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3080646Z return func(*args, **kwargs) 2025-12-04T09:37:34.3081008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 413, in forward 2025-12-04T09:37:34.3081434Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:34.3081608Z 2025-12-04T09:37:34.3081748Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3081976Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3082201Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3082426Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3082652Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3082864Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3083076Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3083297Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3083503Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3083710Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3083908Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3084111Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3084314Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3084548Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:34.3084899Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:34.3085214Z res = mod(**inputs) 2025-12-04T09:37:34.3085564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1280, in forward 2025-12-04T09:37:34.3085921Z outputs = self.bert( 2025-12-04T09:37:34.3086270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1000, in forward 2025-12-04T09:37:34.3086645Z encoder_outputs = self.encoder( 2025-12-04T09:37:34.3087002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 650, in forward 2025-12-04T09:37:34.3087373Z layer_outputs = layer_module( 2025-12-04T09:37:34.3087714Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:34.3088071Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:34.3088433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3088795Z return func(*args, **kwargs) 2025-12-04T09:37:34.3089148Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 558, in forward 2025-12-04T09:37:34.3089524Z self_attention_outputs = self.attention( 2025-12-04T09:37:34.3089921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3090284Z return func(*args, **kwargs) 2025-12-04T09:37:34.3090641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 488, in forward 2025-12-04T09:37:34.3091007Z self_outputs = self.self( 2025-12-04T09:37:34.3091356Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3091743Z return func(*args, **kwargs) 2025-12-04T09:37:34.3092102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 413, in forward 2025-12-04T09:37:34.3092516Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:34.3092698Z 2025-12-04T09:37:34.3092777Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3092983Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3093180Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3093381Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3093578Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3093773Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3093964Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3094164Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3094366Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3094561Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3094761Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3094958Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3095148Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3095378Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:34.3095729Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:34.3096046Z res = mod(**inputs) 2025-12-04T09:37:34.3096389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1280, in forward 2025-12-04T09:37:34.3096764Z outputs = self.bert( 2025-12-04T09:37:34.3097106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1000, in forward 2025-12-04T09:37:34.3097463Z encoder_outputs = self.encoder( 2025-12-04T09:37:34.3097822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 650, in forward 2025-12-04T09:37:34.3098187Z layer_outputs = layer_module( 2025-12-04T09:37:34.3098515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:34.3098852Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:34.3099214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3099565Z return func(*args, **kwargs) 2025-12-04T09:37:34.3099906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 558, in forward 2025-12-04T09:37:34.3100275Z self_attention_outputs = self.attention( 2025-12-04T09:37:34.3100635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3100987Z return func(*args, **kwargs) 2025-12-04T09:37:34.3101325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 488, in forward 2025-12-04T09:37:34.3101681Z self_outputs = self.self( 2025-12-04T09:37:34.3102020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3102365Z return func(*args, **kwargs) 2025-12-04T09:37:34.3102715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 413, in forward 2025-12-04T09:37:34.3103160Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:34.3103331Z 2025-12-04T09:37:34.3103413Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3103606Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3103804Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3103998Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3104929Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3105122Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3105317Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3105546Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3105743Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3105930Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3106127Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3106319Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3106511Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3106728Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:34.3107070Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:34.3107380Z res = mod(**inputs) 2025-12-04T09:37:34.3107719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1280, in forward 2025-12-04T09:37:34.3108078Z outputs = self.bert( 2025-12-04T09:37:34.3108421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1000, in forward 2025-12-04T09:37:34.3108782Z encoder_outputs = self.encoder( 2025-12-04T09:37:34.3109128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 650, in forward 2025-12-04T09:37:34.3109485Z layer_outputs = layer_module( 2025-12-04T09:37:34.3109817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:34.3110154Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:34.3110514Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3110865Z return func(*args, **kwargs) 2025-12-04T09:37:34.3111216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 558, in forward 2025-12-04T09:37:34.3111583Z self_attention_outputs = self.attention( 2025-12-04T09:37:34.3111946Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3112296Z return func(*args, **kwargs) 2025-12-04T09:37:34.3112636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 488, in forward 2025-12-04T09:37:34.3112995Z self_outputs = self.self( 2025-12-04T09:37:34.3113337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3113685Z return func(*args, **kwargs) 2025-12-04T09:37:34.3114023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 413, in forward 2025-12-04T09:37:34.3114432Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:34.3114602Z 2025-12-04T09:37:34.3114683Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3114876Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3115075Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3115272Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3115463Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3115650Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3115842Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3116038Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3116264Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3116463Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3116657Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3116842Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3117036Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3117269Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:34.3117645Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:34.3117944Z res = mod(**inputs) 2025-12-04T09:37:34.3118288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1280, in forward 2025-12-04T09:37:34.3118648Z outputs = self.bert( 2025-12-04T09:37:34.3118984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1000, in forward 2025-12-04T09:37:34.3119347Z encoder_outputs = self.encoder( 2025-12-04T09:37:34.3119705Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 650, in forward 2025-12-04T09:37:34.3120066Z layer_outputs = layer_module( 2025-12-04T09:37:34.3120400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:34.3120754Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:34.3121131Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3121489Z return func(*args, **kwargs) 2025-12-04T09:37:34.3121935Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 558, in forward 2025-12-04T09:37:34.3122338Z self_attention_outputs = self.attention( 2025-12-04T09:37:34.3122727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3123078Z return func(*args, **kwargs) 2025-12-04T09:37:34.3123447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 488, in forward 2025-12-04T09:37:34.3123877Z self_outputs = self.self( 2025-12-04T09:37:34.3124281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3124664Z return func(*args, **kwargs) 2025-12-04T09:37:34.3125035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 413, in forward 2025-12-04T09:37:34.3125473Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:34.3125655Z 2025-12-04T09:37:34.3125736Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3125952Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3126163Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3126363Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3126570Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3126777Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3126982Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3127180Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3127385Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3127591Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3127789Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3127993Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3128197Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3128424Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:34.3128787Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:34.3129110Z res = mod(**inputs) 2025-12-04T09:37:34.3129471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1280, in forward 2025-12-04T09:37:34.3129884Z outputs = self.bert( 2025-12-04T09:37:34.3130405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1000, in forward 2025-12-04T09:37:34.3130801Z encoder_outputs = self.encoder( 2025-12-04T09:37:34.3131170Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 650, in forward 2025-12-04T09:37:34.3131613Z layer_outputs = layer_module( 2025-12-04T09:37:34.3131961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:34.3132323Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:34.3132695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3133065Z return func(*args, **kwargs) 2025-12-04T09:37:34.3133442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 558, in forward 2025-12-04T09:37:34.3133803Z self_attention_outputs = self.attention( 2025-12-04T09:37:34.3134166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3134513Z return func(*args, **kwargs) 2025-12-04T09:37:34.3134860Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 488, in forward 2025-12-04T09:37:34.3135213Z self_outputs = self.self( 2025-12-04T09:37:34.3135551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3135897Z return func(*args, **kwargs) 2025-12-04T09:37:34.3136235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 413, in forward 2025-12-04T09:37:34.3136642Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:34.3136820Z 2025-12-04T09:37:34.3136896Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3137094Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3137290Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3137487Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3137682Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3137872Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3138067Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3138262Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3138454Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3138642Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3138838Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3139033Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3139219Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3139443Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:34.3139787Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:34.3140082Z res = mod(**inputs) 2025-12-04T09:37:34.3140421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1280, in forward 2025-12-04T09:37:34.3140772Z outputs = self.bert( 2025-12-04T09:37:34.3141111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1000, in forward 2025-12-04T09:37:34.3141468Z encoder_outputs = self.encoder( 2025-12-04T09:37:34.3141817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 650, in forward 2025-12-04T09:37:34.3142172Z layer_outputs = layer_module( 2025-12-04T09:37:34.3142493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:37:34.3142834Z return super().__call__(*args, **kwargs) 2025-12-04T09:37:34.3143241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3143599Z return func(*args, **kwargs) 2025-12-04T09:37:34.3143941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 558, in forward 2025-12-04T09:37:34.3144340Z self_attention_outputs = self.attention( 2025-12-04T09:37:34.3144701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3145040Z return func(*args, **kwargs) 2025-12-04T09:37:34.3145385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 488, in forward 2025-12-04T09:37:34.3145743Z self_outputs = self.self( 2025-12-04T09:37:34.3146086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:37:34.3146427Z return func(*args, **kwargs) 2025-12-04T09:37:34.3146775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 413, in forward 2025-12-04T09:37:34.3147184Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:37:34.3147353Z 2025-12-04T09:37:34.3147434Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3147628Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3147823Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3148017Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3148202Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3148396Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3148593Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3148777Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3148970Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3149162Z cudagraph partition due to non gpu ops 2025-12-04T09:37:34.3149376Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:37:34.3149724Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:37:34.3150030Z res = mod(**inputs) 2025-12-04T09:37:34.3150369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1300, in forward 2025-12-04T09:37:34.3150838Z masked_lm_loss = loss_fct(prediction_scores.view(-1, self.config.vocab_size), labels.view(-1)) 2025-12-04T09:37:34.3151072Z 2025-12-04T09:37:43.7203263Z Compilation time (from dynamo_timed): 18.229668181 2025-12-04T09:37:43.7232195Z pass 2025-12-04T09:37:43.7234472Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:37:43.7235412Z TIMING: _recursive_pre_grad_passes:0.02842 _recursive_joint_graph_passes:0.38639 _recursive_post_grad_passes:0.06002 async_compile.wait:0.80826 code_gen:9.06054 inductor_compile:10.96212 backend_compile:15.48695 gc:0.00014 entire_frame_compile:18.22967 total_wall_time:18.22967 2025-12-04T09:37:43.7240137Z STATS: call_* op count: 291 | FakeTensorMode.__torch_dispatch__:20442 | FakeTensor.__torch_dispatch__:2935 | ProxyTorchDispatchMode.__torch_dispatch__:4058 2025-12-04T09:37:43.7240692Z Dynamo produced 1 graphs covering 291 ops with 0 graph breaks (0 unique) 2025-12-04T09:37:46.4525809Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:37:46.4526718Z import pynvml # type: ignore[import] 2025-12-04T09:37:49.6027218Z 2025-12-04T09:38:05.1811732Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:38:05.1815739Z loading model: 0it [00:15, ?it/s] 2025-12-04T09:38:05.1820224Z cpu eval BlenderbotForCausalLM 2025-12-04T09:38:05.1993662Z Compilation time (from dynamo_timed): 0 2025-12-04T09:38:05.1993945Z pass_due_to_skip 2025-12-04T09:38:05.1994224Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:38:05.1994541Z TIMING: total_wall_time:0 2025-12-04T09:38:05.1998789Z STATS: call_* op count: 0 2025-12-04T09:38:05.1999247Z Dynamo produced 0 graphs covering 0 ops with 0 graph breaks (0 unique) 2025-12-04T09:38:07.2284779Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:38:07.2285674Z import pynvml # type: ignore[import] 2025-12-04T09:38:10.4425110Z 2025-12-04T09:38:17.4247066Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:38:17.4247588Z loading model: 0it [00:06, ?it/s] 2025-12-04T09:38:17.4248013Z cpu eval DebertaV2ForMaskedLM 2025-12-04T09:38:17.4436747Z Compilation time (from dynamo_timed): 0 2025-12-04T09:38:17.4439894Z pass_due_to_skip 2025-12-04T09:38:17.4440257Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:38:17.4446294Z TIMING: total_wall_time:0 2025-12-04T09:38:17.4450890Z STATS: call_* op count: 0 2025-12-04T09:38:17.4452916Z Dynamo produced 0 graphs covering 0 ops with 0 graph breaks (0 unique) 2025-12-04T09:38:19.2858486Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:38:19.2860726Z import pynvml # type: ignore[import] 2025-12-04T09:38:22.4902563Z 2025-12-04T09:38:23.1019054Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:38:23.1019625Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:38:23.1020012Z cpu eval DistilBertForMaskedLM 2025-12-04T09:38:23.2520076Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:38:23.3110727Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:38:23.3679484Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:38:28.9243553Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9247420Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9247769Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9248013Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9248246Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9248477Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9248721Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9248946Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9249198Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9249423Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9249654Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9249917Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9250148Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9250426Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:28.9250881Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:38:28.9251256Z res = mod(**inputs) 2025-12-04T09:38:28.9251749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 814, in forward 2025-12-04T09:38:28.9252210Z dlbrt_output = self.distilbert( 2025-12-04T09:38:28.9252656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 724, in forward 2025-12-04T09:38:28.9253492Z return self.transformer( 2025-12-04T09:38:28.9253909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 531, in forward 2025-12-04T09:38:28.9254321Z layer_outputs = layer_module( 2025-12-04T09:38:28.9254686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:28.9255168Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:28.9255593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 466, in forward 2025-12-04T09:38:28.9256022Z sa_output = self.attention( 2025-12-04T09:38:28.9256423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 392, in forward 2025-12-04T09:38:28.9256913Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:38:28.9257120Z 2025-12-04T09:38:28.9257207Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9257419Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9257623Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9257828Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9258036Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9258250Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9258463Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9258668Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9258884Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9259091Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9259296Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9259501Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9259697Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9259932Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:28.9260318Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:38:28.9260643Z res = mod(**inputs) 2025-12-04T09:38:28.9261040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 814, in forward 2025-12-04T09:38:28.9261470Z dlbrt_output = self.distilbert( 2025-12-04T09:38:28.9261886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 724, in forward 2025-12-04T09:38:28.9262316Z return self.transformer( 2025-12-04T09:38:28.9262717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 531, in forward 2025-12-04T09:38:28.9263114Z layer_outputs = layer_module( 2025-12-04T09:38:28.9263451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:28.9263804Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:28.9264209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 466, in forward 2025-12-04T09:38:28.9264601Z sa_output = self.attention( 2025-12-04T09:38:28.9264981Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 392, in forward 2025-12-04T09:38:28.9265545Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:38:28.9265768Z 2025-12-04T09:38:28.9265857Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9266057Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9266266Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9266468Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9266671Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9266865Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9267068Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9267400Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9267602Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9267800Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9267999Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9268191Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9268393Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9268673Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:28.9269021Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:38:28.9269336Z res = mod(**inputs) 2025-12-04T09:38:28.9269714Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 814, in forward 2025-12-04T09:38:28.9270117Z dlbrt_output = self.distilbert( 2025-12-04T09:38:28.9270506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 724, in forward 2025-12-04T09:38:28.9270911Z return self.transformer( 2025-12-04T09:38:28.9271298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 531, in forward 2025-12-04T09:38:28.9271700Z layer_outputs = layer_module( 2025-12-04T09:38:28.9272034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:28.9272396Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:28.9272797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 466, in forward 2025-12-04T09:38:28.9273189Z sa_output = self.attention( 2025-12-04T09:38:28.9273577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 392, in forward 2025-12-04T09:38:28.9274031Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:38:28.9274210Z 2025-12-04T09:38:28.9274297Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9274496Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9274697Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9274897Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9275088Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9275288Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9275511Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9275713Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9275903Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9276101Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9276298Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9276497Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9276686Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9276913Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:28.9277270Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:38:28.9277575Z res = mod(**inputs) 2025-12-04T09:38:28.9277960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 814, in forward 2025-12-04T09:38:28.9278359Z dlbrt_output = self.distilbert( 2025-12-04T09:38:28.9278754Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 724, in forward 2025-12-04T09:38:28.9279147Z return self.transformer( 2025-12-04T09:38:28.9279532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 531, in forward 2025-12-04T09:38:28.9279929Z layer_outputs = layer_module( 2025-12-04T09:38:28.9280262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:28.9280618Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:28.9281055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 466, in forward 2025-12-04T09:38:28.9281455Z sa_output = self.attention( 2025-12-04T09:38:28.9281983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 392, in forward 2025-12-04T09:38:28.9282541Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:38:28.9282738Z 2025-12-04T09:38:28.9282834Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9283058Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9283275Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9283488Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9283699Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9283902Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9284113Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9284337Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9284531Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9284730Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9284930Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9285124Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9285325Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9285561Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:28.9285914Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:38:28.9286225Z res = mod(**inputs) 2025-12-04T09:38:28.9286636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 814, in forward 2025-12-04T09:38:28.9287027Z dlbrt_output = self.distilbert( 2025-12-04T09:38:28.9287412Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 724, in forward 2025-12-04T09:38:28.9287805Z return self.transformer( 2025-12-04T09:38:28.9288183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 531, in forward 2025-12-04T09:38:28.9288606Z layer_outputs = layer_module( 2025-12-04T09:38:28.9288932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:28.9289286Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:28.9289684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 466, in forward 2025-12-04T09:38:28.9290063Z sa_output = self.attention( 2025-12-04T09:38:28.9290440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 392, in forward 2025-12-04T09:38:28.9290881Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:38:28.9291053Z 2025-12-04T09:38:28.9291138Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9291333Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9291530Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9291727Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9291915Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9292114Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9292312Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9292508Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9292695Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9292888Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9293078Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9293262Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9293454Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9293675Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:28.9294049Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:38:28.9294356Z res = mod(**inputs) 2025-12-04T09:38:28.9294728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 814, in forward 2025-12-04T09:38:28.9295126Z dlbrt_output = self.distilbert( 2025-12-04T09:38:28.9295557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 724, in forward 2025-12-04T09:38:28.9295946Z return self.transformer( 2025-12-04T09:38:28.9296321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 531, in forward 2025-12-04T09:38:28.9296701Z layer_outputs = layer_module( 2025-12-04T09:38:28.9297031Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:28.9297381Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:28.9297772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 466, in forward 2025-12-04T09:38:28.9298151Z sa_output = self.attention( 2025-12-04T09:38:28.9298531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 392, in forward 2025-12-04T09:38:28.9298976Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:38:28.9299146Z 2025-12-04T09:38:28.9299229Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9299423Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9299625Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9299821Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9300010Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9300207Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9300404Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9300592Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9300787Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9300982Z cudagraph partition due to non gpu ops 2025-12-04T09:38:28.9301199Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:28.9301541Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:38:28.9301851Z res = mod(**inputs) 2025-12-04T09:38:28.9302225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 831, in forward 2025-12-04T09:38:28.9302743Z mlm_loss = self.mlm_loss_fct(prediction_logits.view(-1, prediction_logits.size(-1)), labels.view(-1)) 2025-12-04T09:38:28.9302989Z 2025-12-04T09:38:37.4271949Z Compilation time (from dynamo_timed): 13.574297892 2025-12-04T09:38:37.4272555Z pass 2025-12-04T09:38:37.4273066Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:38:37.4274491Z TIMING: _recursive_pre_grad_passes:0.01681 _recursive_joint_graph_passes:0.25756 _recursive_post_grad_passes:0.0403 async_compile.wait:0.77387 code_gen:8.33089 inductor_compile:9.46517 backend_compile:11.96213 gc:0.00017 entire_frame_compile:13.5743 total_wall_time:13.5743 2025-12-04T09:38:37.4275600Z STATS: call_* op count: 155 | FakeTensorMode.__torch_dispatch__:10973 | FakeTensor.__torch_dispatch__:1621 | ProxyTorchDispatchMode.__torch_dispatch__:2149 2025-12-04T09:38:37.4276276Z Dynamo produced 1 graphs covering 155 ops with 0 graph breaks (0 unique) 2025-12-04T09:38:39.6637997Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:38:39.6638795Z import pynvml # type: ignore[import] 2025-12-04T09:38:42.9038438Z 2025-12-04T09:38:43.7863133Z loading model: 0it [00:00, ?it/s]`loss_type=None` was set in the config but it is unrecognized. Using the default loss: `ForCausalLMLoss`. 2025-12-04T09:38:43.7865554Z WARNING:transformers.modeling_utils:`loss_type=None` was set in the config but it is unrecognized. Using the default loss: `ForCausalLMLoss`. 2025-12-04T09:38:43.8156401Z 2025-12-04T09:38:43.8157269Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:38:43.8160503Z cpu eval DistillGPT2 2025-12-04T09:38:44.2313472Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:38:44.3609831Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:38:44.4919651Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:38:51.4026348Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4026944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4027366Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4027761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 873, in forward 2025-12-04T09:38:51.4028146Z causal_mask = create_causal_mask( 2025-12-04T09:38:51.4028532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 788, in create_causal_mask 2025-12-04T09:38:51.4029030Z early_exit, attention_mask, packed_sequence_mask, kv_length, kv_offset = _preprocess_mask_arguments( 2025-12-04T09:38:51.4029536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 740, in _preprocess_mask_arguments 2025-12-04T09:38:51.4029980Z packed_sequence_mask = find_packed_sequence_indices(position_ids) 2025-12-04T09:38:51.4030571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 655, in find_packed_sequence_indices 2025-12-04T09:38:51.4031098Z first_dummy_value = position_ids[:, :1] - 1 # We just need the diff on this first value to be 1 2025-12-04T09:38:51.4031322Z 2025-12-04T09:38:51.4031415Z cudagraph partition due to non gpu ops 2025-12-04T09:38:51.4031653Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4032068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4032463Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4032846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4033245Z outputs = block( 2025-12-04T09:38:51.4033570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4033934Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4034308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4034669Z return func(*args, **kwargs) 2025-12-04T09:38:51.4035038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4035433Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4035809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4036176Z return func(*args, **kwargs) 2025-12-04T09:38:51.4036541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:38:51.4037024Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:38:51.4037807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4038215Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4038390Z 2025-12-04T09:38:51.4038511Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4038920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4039391Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4039780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4040170Z outputs = block( 2025-12-04T09:38:51.4040481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4040841Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4041251Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4041693Z return func(*args, **kwargs) 2025-12-04T09:38:51.4042139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4042592Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4043031Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4043419Z return func(*args, **kwargs) 2025-12-04T09:38:51.4043795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:38:51.4044270Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:38:51.4044720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4045164Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4045346Z 2025-12-04T09:38:51.4045451Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4045851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4046235Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4046628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 873, in forward 2025-12-04T09:38:51.4047016Z causal_mask = create_causal_mask( 2025-12-04T09:38:51.4047390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 788, in create_causal_mask 2025-12-04T09:38:51.4047886Z early_exit, attention_mask, packed_sequence_mask, kv_length, kv_offset = _preprocess_mask_arguments( 2025-12-04T09:38:51.4048424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 740, in _preprocess_mask_arguments 2025-12-04T09:38:51.4048876Z packed_sequence_mask = find_packed_sequence_indices(position_ids) 2025-12-04T09:38:51.4049327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 656, in find_packed_sequence_indices 2025-12-04T09:38:51.4049805Z position_diff = torch.diff(position_ids, prepend=first_dummy_value, dim=-1) 2025-12-04T09:38:51.4050030Z 2025-12-04T09:38:51.4050114Z cudagraph partition due to non gpu ops 2025-12-04T09:38:51.4050355Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4050778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4051179Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4051614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 873, in forward 2025-12-04T09:38:51.4051998Z causal_mask = create_causal_mask( 2025-12-04T09:38:51.4052372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 788, in create_causal_mask 2025-12-04T09:38:51.4052855Z early_exit, attention_mask, packed_sequence_mask, kv_length, kv_offset = _preprocess_mask_arguments( 2025-12-04T09:38:51.4053405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 740, in _preprocess_mask_arguments 2025-12-04T09:38:51.4053839Z packed_sequence_mask = find_packed_sequence_indices(position_ids) 2025-12-04T09:38:51.4054292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 657, in find_packed_sequence_indices 2025-12-04T09:38:51.4054706Z packed_sequence_mask = (position_diff != 1).cumsum(-1) 2025-12-04T09:38:51.4054873Z 2025-12-04T09:38:51.4054975Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4055381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4055763Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4056133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 873, in forward 2025-12-04T09:38:51.4056508Z causal_mask = create_causal_mask( 2025-12-04T09:38:51.4056866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 788, in create_causal_mask 2025-12-04T09:38:51.4057335Z early_exit, attention_mask, packed_sequence_mask, kv_length, kv_offset = _preprocess_mask_arguments( 2025-12-04T09:38:51.4057838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 740, in _preprocess_mask_arguments 2025-12-04T09:38:51.4058280Z packed_sequence_mask = find_packed_sequence_indices(position_ids) 2025-12-04T09:38:51.4058731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 657, in find_packed_sequence_indices 2025-12-04T09:38:51.4059152Z packed_sequence_mask = (position_diff != 1).cumsum(-1) 2025-12-04T09:38:51.4059322Z 2025-12-04T09:38:51.4059427Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4059839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4060233Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4060611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4060980Z outputs = block( 2025-12-04T09:38:51.4061317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4061673Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4062044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4062407Z return func(*args, **kwargs) 2025-12-04T09:38:51.4062771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4063159Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4063576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4064108Z return func(*args, **kwargs) 2025-12-04T09:38:51.4064482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:38:51.4064873Z attn_output, attn_weights = attention_interface( 2025-12-04T09:38:51.4065299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:38:51.4065814Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:38:51.4065994Z 2025-12-04T09:38:51.4066082Z cudagraph partition due to non gpu ops 2025-12-04T09:38:51.4066318Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4066715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4067126Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4067503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4067859Z outputs = block( 2025-12-04T09:38:51.4068175Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4068537Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4068897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4069243Z return func(*args, **kwargs) 2025-12-04T09:38:51.4069592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4069969Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4070327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4070680Z return func(*args, **kwargs) 2025-12-04T09:38:51.4071024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:38:51.4071403Z attn_output, attn_weights = attention_interface( 2025-12-04T09:38:51.4071812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:38:51.4072264Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:38:51.4072434Z 2025-12-04T09:38:51.4072540Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4072928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4073304Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4073679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4074037Z outputs = block( 2025-12-04T09:38:51.4074344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4074695Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4075065Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4075415Z return func(*args, **kwargs) 2025-12-04T09:38:51.4075761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4076139Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4076505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4076856Z return func(*args, **kwargs) 2025-12-04T09:38:51.4077196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:38:51.4077565Z attn_output = self.c_proj(attn_output) 2025-12-04T09:38:51.4077906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4078279Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4078452Z 2025-12-04T09:38:51.4078550Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4079003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4079380Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4079741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4080125Z outputs = block( 2025-12-04T09:38:51.4080430Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4080763Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4081123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4081506Z return func(*args, **kwargs) 2025-12-04T09:38:51.4082043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4082483Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4082911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4083279Z return func(*args, **kwargs) 2025-12-04T09:38:51.4083694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:38:51.4084170Z attn_output = self.c_proj(attn_output) 2025-12-04T09:38:51.4084569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4085021Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4085211Z 2025-12-04T09:38:51.4085325Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4085785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4086240Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4086673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4087078Z outputs = block( 2025-12-04T09:38:51.4087429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4087833Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4088261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4088677Z return func(*args, **kwargs) 2025-12-04T09:38:51.4089028Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4089421Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4089807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:38:51.4090186Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:38:51.4090534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4090914Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4091079Z 2025-12-04T09:38:51.4091176Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4091576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4091959Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4092334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4092691Z outputs = block( 2025-12-04T09:38:51.4093045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4093392Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4093741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4094098Z return func(*args, **kwargs) 2025-12-04T09:38:51.4094488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4094886Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4095268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:38:51.4095641Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:38:51.4095986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4096364Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4096535Z 2025-12-04T09:38:51.4096634Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4097032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4097408Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4097775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4098133Z outputs = block( 2025-12-04T09:38:51.4098440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4098784Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4099141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4099500Z return func(*args, **kwargs) 2025-12-04T09:38:51.4099854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4100239Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4100624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:38:51.4100995Z hidden_states = self.act(hidden_states) 2025-12-04T09:38:51.4101334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:38:51.4101762Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:38:51.4101991Z 2025-12-04T09:38:51.4102091Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4102489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4102870Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4103234Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4103592Z outputs = block( 2025-12-04T09:38:51.4103899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4104237Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4104597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4104953Z return func(*args, **kwargs) 2025-12-04T09:38:51.4105304Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4105686Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4106101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:38:51.4106478Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:38:51.4106828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4107205Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4107408Z 2025-12-04T09:38:51.4107507Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4107899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4108265Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4108628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4108982Z outputs = block( 2025-12-04T09:38:51.4109290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4109624Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4109983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4110333Z return func(*args, **kwargs) 2025-12-04T09:38:51.4110678Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4111065Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4111447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:38:51.4111820Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:38:51.4112155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4112532Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4112701Z 2025-12-04T09:38:51.4112801Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4113192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4113556Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4113924Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4114277Z outputs = block( 2025-12-04T09:38:51.4114572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4114910Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4115267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4115616Z return func(*args, **kwargs) 2025-12-04T09:38:51.4115961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4116334Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4116700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4117053Z return func(*args, **kwargs) 2025-12-04T09:38:51.4117395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:38:51.4117859Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:38:51.4118291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4118659Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4118825Z 2025-12-04T09:38:51.4118955Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4119349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4119726Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4120085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4120481Z outputs = block( 2025-12-04T09:38:51.4120786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4121126Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4121490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4121949Z return func(*args, **kwargs) 2025-12-04T09:38:51.4122326Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4122719Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4123108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4123480Z return func(*args, **kwargs) 2025-12-04T09:38:51.4123851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:38:51.4124342Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:38:51.4124808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4125205Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4125373Z 2025-12-04T09:38:51.4125466Z cudagraph partition due to non gpu ops 2025-12-04T09:38:51.4125703Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4126119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4126517Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4126897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4127272Z outputs = block( 2025-12-04T09:38:51.4127593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4127953Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4128327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4128698Z return func(*args, **kwargs) 2025-12-04T09:38:51.4129070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4129457Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4129845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4130390Z return func(*args, **kwargs) 2025-12-04T09:38:51.4130766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:38:51.4131174Z attn_output, attn_weights = attention_interface( 2025-12-04T09:38:51.4131604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:38:51.4132061Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:38:51.4132233Z 2025-12-04T09:38:51.4132341Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4132819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4133209Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4133581Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4133937Z outputs = block( 2025-12-04T09:38:51.4134239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4134636Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4134996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4135336Z return func(*args, **kwargs) 2025-12-04T09:38:51.4135681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4136053Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4136420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4136763Z return func(*args, **kwargs) 2025-12-04T09:38:51.4137110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:38:51.4137479Z attn_output = self.c_proj(attn_output) 2025-12-04T09:38:51.4137812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4138185Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4138352Z 2025-12-04T09:38:51.4138451Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4138846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4139214Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4139579Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4139927Z outputs = block( 2025-12-04T09:38:51.4140233Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4140568Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4140923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4141273Z return func(*args, **kwargs) 2025-12-04T09:38:51.4141612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4141982Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4142345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4142697Z return func(*args, **kwargs) 2025-12-04T09:38:51.4143037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:38:51.4143403Z attn_output = self.c_proj(attn_output) 2025-12-04T09:38:51.4143738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4144105Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4144271Z 2025-12-04T09:38:51.4144370Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4144757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4145128Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4145483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4146558Z outputs = block( 2025-12-04T09:38:51.4146890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4147232Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4147585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4147972Z return func(*args, **kwargs) 2025-12-04T09:38:51.4148323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4148706Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4149091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:38:51.4149459Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:38:51.4149795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4150160Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4150325Z 2025-12-04T09:38:51.4150423Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4150813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4151188Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4151544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4151892Z outputs = block( 2025-12-04T09:38:51.4152199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4152535Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4152898Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4153250Z return func(*args, **kwargs) 2025-12-04T09:38:51.4153599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4153979Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4154364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:38:51.4154730Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:38:51.4155059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4155432Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4155597Z 2025-12-04T09:38:51.4155696Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4156089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4156459Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4156823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4157173Z outputs = block( 2025-12-04T09:38:51.4157478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4157810Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4158167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4158516Z return func(*args, **kwargs) 2025-12-04T09:38:51.4158858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4159245Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4159655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:38:51.4160025Z hidden_states = self.act(hidden_states) 2025-12-04T09:38:51.4160357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:38:51.4160820Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:38:51.4161043Z 2025-12-04T09:38:51.4161143Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4161628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4162028Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4162415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4162794Z outputs = block( 2025-12-04T09:38:51.4163103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4163461Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4163825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4164179Z return func(*args, **kwargs) 2025-12-04T09:38:51.4164540Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4164948Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4165356Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:38:51.4165749Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:38:51.4166100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4166491Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4166660Z 2025-12-04T09:38:51.4166770Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4167175Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4167575Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4167958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4168327Z outputs = block( 2025-12-04T09:38:51.4168639Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4169003Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4169382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4169753Z return func(*args, **kwargs) 2025-12-04T09:38:51.4170114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4170522Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4170924Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:38:51.4171312Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:38:51.4171670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4172067Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4172232Z 2025-12-04T09:38:51.4172343Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4172784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4173177Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4173559Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4173928Z outputs = block( 2025-12-04T09:38:51.4174268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4174610Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4174965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4175308Z return func(*args, **kwargs) 2025-12-04T09:38:51.4175659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4176030Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4176398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4176739Z return func(*args, **kwargs) 2025-12-04T09:38:51.4177086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:38:51.4177552Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:38:51.4177986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4178351Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4178518Z 2025-12-04T09:38:51.4178615Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4179007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4179378Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4179744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4180096Z outputs = block( 2025-12-04T09:38:51.4180401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4180739Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4181099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4181454Z return func(*args, **kwargs) 2025-12-04T09:38:51.4181798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4182170Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4182541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4182889Z return func(*args, **kwargs) 2025-12-04T09:38:51.4183229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:38:51.4183691Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:38:51.4184124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4184495Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4184655Z 2025-12-04T09:38:51.4184735Z cudagraph partition due to non gpu ops 2025-12-04T09:38:51.4184966Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4185359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4185755Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4186122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4186472Z outputs = block( 2025-12-04T09:38:51.4186776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4187141Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4187502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4187856Z return func(*args, **kwargs) 2025-12-04T09:38:51.4188203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4188569Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4188939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4189287Z return func(*args, **kwargs) 2025-12-04T09:38:51.4189626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:38:51.4190007Z attn_output, attn_weights = attention_interface( 2025-12-04T09:38:51.4190426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:38:51.4190877Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:38:51.4191045Z 2025-12-04T09:38:51.4191144Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4191534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4191903Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4192267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4192609Z outputs = block( 2025-12-04T09:38:51.4192912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4193249Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4193600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4193950Z return func(*args, **kwargs) 2025-12-04T09:38:51.4194291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4194660Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4195020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4195372Z return func(*args, **kwargs) 2025-12-04T09:38:51.4195719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:38:51.4196082Z attn_output = self.c_proj(attn_output) 2025-12-04T09:38:51.4196421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4196799Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4196963Z 2025-12-04T09:38:51.4197068Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4197466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4197844Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4198216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4198566Z outputs = block( 2025-12-04T09:38:51.4198985Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4199338Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4199698Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4200045Z return func(*args, **kwargs) 2025-12-04T09:38:51.4200426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4200800Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4201177Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4201626Z return func(*args, **kwargs) 2025-12-04T09:38:51.4202019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:38:51.4202423Z attn_output = self.c_proj(attn_output) 2025-12-04T09:38:51.4202791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4203235Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4203406Z 2025-12-04T09:38:51.4203507Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4203908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4204276Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4204641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4204996Z outputs = block( 2025-12-04T09:38:51.4205317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4205657Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4206017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4206368Z return func(*args, **kwargs) 2025-12-04T09:38:51.4206708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4207096Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4207481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:38:51.4207846Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:38:51.4208176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4208551Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4208719Z 2025-12-04T09:38:51.4208823Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4209215Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4209581Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4209945Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4210302Z outputs = block( 2025-12-04T09:38:51.4210602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4210938Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4211294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4211643Z return func(*args, **kwargs) 2025-12-04T09:38:51.4212019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4212411Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4212792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:38:51.4213160Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:38:51.4213524Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4213899Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4214057Z 2025-12-04T09:38:51.4214164Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4214548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4214921Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4215288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4215642Z outputs = block( 2025-12-04T09:38:51.4215940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4216282Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4216641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4216994Z return func(*args, **kwargs) 2025-12-04T09:38:51.4217332Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4217718Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4218097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:38:51.4218454Z hidden_states = self.act(hidden_states) 2025-12-04T09:38:51.4218789Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:38:51.4219217Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:38:51.4219433Z 2025-12-04T09:38:51.4219539Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4219929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4220300Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4220664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4221015Z outputs = block( 2025-12-04T09:38:51.4221315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4221656Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4222018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4222364Z return func(*args, **kwargs) 2025-12-04T09:38:51.4222711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4223097Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4223476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:38:51.4223843Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:38:51.4224183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4224555Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4224713Z 2025-12-04T09:38:51.4224861Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4225254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4225627Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4225989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4226366Z outputs = block( 2025-12-04T09:38:51.4226673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4227017Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4227379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4227727Z return func(*args, **kwargs) 2025-12-04T09:38:51.4228079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4228472Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4228852Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:38:51.4229228Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:38:51.4229571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4229947Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4230108Z 2025-12-04T09:38:51.4230333Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4230737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4231112Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4231484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4231832Z outputs = block( 2025-12-04T09:38:51.4232142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4232490Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4232843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4233204Z return func(*args, **kwargs) 2025-12-04T09:38:51.4233550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 451, in forward 2025-12-04T09:38:51.4233942Z hidden_states = residual + feed_forward_hidden_states 2025-12-04T09:38:51.4234093Z 2025-12-04T09:38:51.4234191Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4234585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4234954Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4235315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4235659Z outputs = block( 2025-12-04T09:38:51.4235966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4236308Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4236655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4237007Z return func(*args, **kwargs) 2025-12-04T09:38:51.4237353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4237724Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4238154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4238511Z return func(*args, **kwargs) 2025-12-04T09:38:51.4238865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:38:51.4239329Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:38:51.4239822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4240201Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4240362Z 2025-12-04T09:38:51.4240466Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4240850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4241227Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4241666Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4242056Z outputs = block( 2025-12-04T09:38:51.4242381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4242757Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4243124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4243483Z return func(*args, **kwargs) 2025-12-04T09:38:51.4243834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4244211Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4244590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4244943Z return func(*args, **kwargs) 2025-12-04T09:38:51.4245302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:38:51.4245776Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:38:51.4246221Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4246601Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4246774Z 2025-12-04T09:38:51.4246858Z cudagraph partition due to non gpu ops 2025-12-04T09:38:51.4247093Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4247492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4247880Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4248258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4248623Z outputs = block( 2025-12-04T09:38:51.4248932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4249286Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4249658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4250018Z return func(*args, **kwargs) 2025-12-04T09:38:51.4250367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4250750Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4251127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4251514Z return func(*args, **kwargs) 2025-12-04T09:38:51.4251870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:38:51.4252260Z attn_output, attn_weights = attention_interface( 2025-12-04T09:38:51.4252696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:38:51.4253190Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:38:51.4253374Z 2025-12-04T09:38:51.4253474Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4253873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4254258Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4254627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4254986Z outputs = block( 2025-12-04T09:38:51.4255301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4255644Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4256015Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4256379Z return func(*args, **kwargs) 2025-12-04T09:38:51.4256739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4257114Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4257502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4257852Z return func(*args, **kwargs) 2025-12-04T09:38:51.4258191Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:38:51.4258562Z attn_output = self.c_proj(attn_output) 2025-12-04T09:38:51.4258899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4259274Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4259435Z 2025-12-04T09:38:51.4259534Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4259930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4260300Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4260660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4261003Z outputs = block( 2025-12-04T09:38:51.4261311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4261694Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4262050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4262401Z return func(*args, **kwargs) 2025-12-04T09:38:51.4262759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4263135Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4263497Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4263850Z return func(*args, **kwargs) 2025-12-04T09:38:51.4264197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:38:51.4264569Z attn_output = self.c_proj(attn_output) 2025-12-04T09:38:51.4264937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4265312Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4265472Z 2025-12-04T09:38:51.4265581Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4265998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4266378Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4266746Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4267097Z outputs = block( 2025-12-04T09:38:51.4267397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4267743Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4268105Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4268453Z return func(*args, **kwargs) 2025-12-04T09:38:51.4268809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4269200Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4269586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:38:51.4269948Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:38:51.4270287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4270659Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4270818Z 2025-12-04T09:38:51.4270925Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4271316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4271697Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4272065Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4272416Z outputs = block( 2025-12-04T09:38:51.4272725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4273068Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4273427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4273771Z return func(*args, **kwargs) 2025-12-04T09:38:51.4274120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4274513Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4274894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:38:51.4275254Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:38:51.4275592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4275969Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4276128Z 2025-12-04T09:38:51.4276224Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4276617Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4276990Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4277400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4277750Z outputs = block( 2025-12-04T09:38:51.4278055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4278407Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4278827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4279200Z return func(*args, **kwargs) 2025-12-04T09:38:51.4279555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4279952Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4280341Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:38:51.4280720Z hidden_states = self.act(hidden_states) 2025-12-04T09:38:51.4281062Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:38:51.4281502Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:38:51.4281814Z 2025-12-04T09:38:51.4281926Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4282364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4282768Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4283144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4283500Z outputs = block( 2025-12-04T09:38:51.4283819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4284171Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4284535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4284897Z return func(*args, **kwargs) 2025-12-04T09:38:51.4285259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4285660Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4286044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:38:51.4286426Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:38:51.4286779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4287164Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4287329Z 2025-12-04T09:38:51.4287429Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4287832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4288213Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4288580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4288940Z outputs = block( 2025-12-04T09:38:51.4289253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4289605Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4289967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4290329Z return func(*args, **kwargs) 2025-12-04T09:38:51.4290688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4291113Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4291503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:38:51.4291881Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:38:51.4292227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4292666Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4292832Z 2025-12-04T09:38:51.4292932Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4293327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4293703Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4294066Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4294423Z outputs = block( 2025-12-04T09:38:51.4294732Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4295072Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4295437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4295800Z return func(*args, **kwargs) 2025-12-04T09:38:51.4296153Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4296527Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4296900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4297256Z return func(*args, **kwargs) 2025-12-04T09:38:51.4297611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:38:51.4298072Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:38:51.4298510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4298892Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4299054Z 2025-12-04T09:38:51.4299156Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4299552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4299937Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4300310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4300662Z outputs = block( 2025-12-04T09:38:51.4300979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4301328Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4301690Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4302042Z return func(*args, **kwargs) 2025-12-04T09:38:51.4302395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4302775Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4303138Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4303494Z return func(*args, **kwargs) 2025-12-04T09:38:51.4303849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:38:51.4304350Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:38:51.4304781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4305163Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4305354Z 2025-12-04T09:38:51.4305440Z cudagraph partition due to non gpu ops 2025-12-04T09:38:51.4305673Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4306061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4306434Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4306800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4307143Z outputs = block( 2025-12-04T09:38:51.4307451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4307792Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4308149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4308492Z return func(*args, **kwargs) 2025-12-04T09:38:51.4308842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4309214Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4309574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4309922Z return func(*args, **kwargs) 2025-12-04T09:38:51.4310266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:38:51.4310648Z attn_output, attn_weights = attention_interface( 2025-12-04T09:38:51.4311066Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:38:51.4311522Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:38:51.4311699Z 2025-12-04T09:38:51.4311798Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4312192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4312558Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4312925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4313278Z outputs = block( 2025-12-04T09:38:51.4313577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4313922Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4314279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4314630Z return func(*args, **kwargs) 2025-12-04T09:38:51.4314971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4315348Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4315719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4316067Z return func(*args, **kwargs) 2025-12-04T09:38:51.4316407Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:38:51.4316775Z attn_output = self.c_proj(attn_output) 2025-12-04T09:38:51.4317143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4317517Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4317687Z 2025-12-04T09:38:51.4317787Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4318181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4318595Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4318958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4319307Z outputs = block( 2025-12-04T09:38:51.4319617Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4319963Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4320326Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4320686Z return func(*args, **kwargs) 2025-12-04T09:38:51.4321051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4321429Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4321888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4322251Z return func(*args, **kwargs) 2025-12-04T09:38:51.4322614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:38:51.4322987Z attn_output = self.c_proj(attn_output) 2025-12-04T09:38:51.4323336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4323720Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4323892Z 2025-12-04T09:38:51.4323994Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4324399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4324782Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4325159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4325513Z outputs = block( 2025-12-04T09:38:51.4325825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4326178Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4326542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4326892Z return func(*args, **kwargs) 2025-12-04T09:38:51.4327250Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4327648Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4328033Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:38:51.4328414Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:38:51.4328756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4329136Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4329297Z 2025-12-04T09:38:51.4329396Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4329799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4330285Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4330748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4331109Z outputs = block( 2025-12-04T09:38:51.4331426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4331780Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4332199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4332561Z return func(*args, **kwargs) 2025-12-04T09:38:51.4332921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4333320Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4333712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:38:51.4334092Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:38:51.4334439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4334818Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4334987Z 2025-12-04T09:38:51.4335088Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4335497Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4335878Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4336243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4336604Z outputs = block( 2025-12-04T09:38:51.4336942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4337282Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4337631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4337980Z return func(*args, **kwargs) 2025-12-04T09:38:51.4338329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4338709Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4339094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:38:51.4339460Z hidden_states = self.act(hidden_states) 2025-12-04T09:38:51.4339793Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:38:51.4340213Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:38:51.4340436Z 2025-12-04T09:38:51.4340536Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4340928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4341304Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4341662Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4342015Z outputs = block( 2025-12-04T09:38:51.4342322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4342655Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4343010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4343359Z return func(*args, **kwargs) 2025-12-04T09:38:51.4343737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4344123Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4344509Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:38:51.4344910Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:38:51.4345248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4345614Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4345778Z 2025-12-04T09:38:51.4345878Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4346267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4346632Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4346997Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4347346Z outputs = block( 2025-12-04T09:38:51.4347651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4347982Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4348344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4348700Z return func(*args, **kwargs) 2025-12-04T09:38:51.4349052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4349434Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4349820Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:38:51.4350194Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:38:51.4350531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4350911Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4351079Z 2025-12-04T09:38:51.4351178Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4351571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4351940Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4352305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4352662Z outputs = block( 2025-12-04T09:38:51.4352958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4353299Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4353655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4354006Z return func(*args, **kwargs) 2025-12-04T09:38:51.4354344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 451, in forward 2025-12-04T09:38:51.4354741Z hidden_states = residual + feed_forward_hidden_states 2025-12-04T09:38:51.4354899Z 2025-12-04T09:38:51.4354997Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4355386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4355752Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4356116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4356495Z outputs = block( 2025-12-04T09:38:51.4356797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4357137Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4357493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4357881Z return func(*args, **kwargs) 2025-12-04T09:38:51.4358226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4358606Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4358980Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4359332Z return func(*args, **kwargs) 2025-12-04T09:38:51.4359675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:38:51.4360142Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:38:51.4360582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4360953Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4361123Z 2025-12-04T09:38:51.4361221Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4361686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4362072Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4362462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4362852Z outputs = block( 2025-12-04T09:38:51.4363181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4363543Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4363916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4364268Z return func(*args, **kwargs) 2025-12-04T09:38:51.4364641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4365041Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4365437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4365822Z return func(*args, **kwargs) 2025-12-04T09:38:51.4366196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:38:51.4366690Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:38:51.4367159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4367561Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4367733Z 2025-12-04T09:38:51.4367831Z cudagraph partition due to non gpu ops 2025-12-04T09:38:51.4368074Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4368498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4368905Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4369290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4369673Z outputs = block( 2025-12-04T09:38:51.4370040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4370413Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4370793Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4371173Z return func(*args, **kwargs) 2025-12-04T09:38:51.4371600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4372001Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4372401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4372784Z return func(*args, **kwargs) 2025-12-04T09:38:51.4373159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:38:51.4373569Z attn_output, attn_weights = attention_interface( 2025-12-04T09:38:51.4374022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:38:51.4374478Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:38:51.4374652Z 2025-12-04T09:38:51.4374758Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4375150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4375522Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4375887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4376230Z outputs = block( 2025-12-04T09:38:51.4376538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4376883Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4377240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4377583Z return func(*args, **kwargs) 2025-12-04T09:38:51.4377932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4378310Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4378675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4379016Z return func(*args, **kwargs) 2025-12-04T09:38:51.4379363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:38:51.4379728Z attn_output = self.c_proj(attn_output) 2025-12-04T09:38:51.4380061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4380436Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4380604Z 2025-12-04T09:38:51.4380706Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4381109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4381479Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4381843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4382196Z outputs = block( 2025-12-04T09:38:51.4382506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4382844Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4383229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4383580Z return func(*args, **kwargs) 2025-12-04T09:38:51.4383921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:38:51.4384291Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:38:51.4384655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4385033Z return func(*args, **kwargs) 2025-12-04T09:38:51.4385374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:38:51.4385744Z attn_output = self.c_proj(attn_output) 2025-12-04T09:38:51.4386089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4386462Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4386630Z 2025-12-04T09:38:51.4386735Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4387135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4387511Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4387872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4388232Z outputs = block( 2025-12-04T09:38:51.4388544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4388889Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4389240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4389599Z return func(*args, **kwargs) 2025-12-04T09:38:51.4389951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4390338Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4390731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:38:51.4391108Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:38:51.4391455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4391826Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4391995Z 2025-12-04T09:38:51.4392094Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4392498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4392875Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4393240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4393598Z outputs = block( 2025-12-04T09:38:51.4393911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4394254Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4394621Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4394974Z return func(*args, **kwargs) 2025-12-04T09:38:51.4395327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4395708Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4396096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:38:51.4396498Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:38:51.4396826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4397197Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4397362Z 2025-12-04T09:38:51.4397460Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4397903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4398268Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4398634Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4398992Z outputs = block( 2025-12-04T09:38:51.4399299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4399636Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4399995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4400348Z return func(*args, **kwargs) 2025-12-04T09:38:51.4400692Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4401087Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4401472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:38:51.4401954Z hidden_states = self.act(hidden_states) 2025-12-04T09:38:51.4402324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:38:51.4402776Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:38:51.4403020Z 2025-12-04T09:38:51.4403127Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4403541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4403929Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4404304Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4404371Z outputs = block( 2025-12-04T09:38:51.4404593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4404673Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4404934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4405005Z return func(*args, **kwargs) 2025-12-04T09:38:51.4405257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4405367Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4405615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:38:51.4405702Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:38:51.4405932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4406049Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4406053Z 2025-12-04T09:38:51.4406164Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:38:51.4406421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1068, in forward 2025-12-04T09:38:51.4406502Z transformer_outputs = self.transformer( 2025-12-04T09:38:51.4406792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:38:51.4406857Z outputs = block( 2025-12-04T09:38:51.4407086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:38:51.4407167Z return super().__call__(*args, **kwargs) 2025-12-04T09:38:51.4407439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:38:51.4407518Z return func(*args, **kwargs) 2025-12-04T09:38:51.4407762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:38:51.4407863Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:38:51.4408112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:38:51.4408199Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:38:51.4408420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:38:51.4408537Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:38:51.4408540Z 2025-12-04T09:38:51.4408621Z cudagraph partition due to non gpu ops 2025-12-04T09:38:51.4408712Z cudagraph partition due to non gpu ops 2025-12-04T09:39:00.3442663Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:00.3443231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/loss/loss_utils.py", line 67, in ForCausalLMLoss 2025-12-04T09:39:00.3443741Z loss = fixed_cross_entropy(logits, shift_labels, num_items_in_batch, ignore_index, **kwargs) 2025-12-04T09:39:00.3444275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/loss/loss_utils.py", line 36, in fixed_cross_entropy 2025-12-04T09:39:00.3444867Z loss = nn.functional.cross_entropy(source, target, ignore_index=ignore_index, reduction=reduction) 2025-12-04T09:39:00.3445202Z 2025-12-04T09:39:01.4487756Z Compilation time (from dynamo_timed): 16.319030345 2025-12-04T09:39:01.4614086Z pass 2025-12-04T09:39:01.4617916Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:39:01.4623242Z TIMING: _recursive_pre_grad_passes:0.02388 _recursive_joint_graph_passes:0.22847 inductor_compile:10.72285 backend_compile:12.52824 gc:0.00193 entire_frame_compile:16.31903 _recursive_post_grad_passes:0.04415 async_compile.wait:1.47126 code_gen:9.53588 total_wall_time:16.31903 2025-12-04T09:39:01.4628002Z STATS: call_* op count: 313 | FakeTensorMode.__torch_dispatch__:10792 | FakeTensor.__torch_dispatch__:1968 | ProxyTorchDispatchMode.__torch_dispatch__:1243 2025-12-04T09:39:01.4629974Z Dynamo produced 3 graphs covering 313 ops with 2 graph breaks (1 unique) 2025-12-04T09:39:03.6356619Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:39:03.6357945Z import pynvml # type: ignore[import] 2025-12-04T09:39:06.7529598Z 2025-12-04T09:39:06.7548865Z loading model: 0it [00:00, ?it/s]If you want to use `ElectraForCausalLM` as a standalone, add `is_decoder=True.` 2025-12-04T09:39:06.7550640Z WARNING:transformers.models.electra.modeling_electra:If you want to use `ElectraForCausalLM` as a standalone, add `is_decoder=True.` 2025-12-04T09:39:06.9489582Z 2025-12-04T09:39:06.9490465Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:39:06.9490916Z cpu eval ElectraForCausalLM 2025-12-04T09:39:07.1035154Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:39:07.1887408Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:39:07.2760921Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:39:16.8794875Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8795203Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8795413Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8796011Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8796226Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8796416Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8796645Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8796847Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8797043Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8797231Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8797438Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8797642Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8797848Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8798048Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8798248Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8798441Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8798643Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8798844Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8799041Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8799243Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8799443Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8799644Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8799838Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8800041Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8800244Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8800434Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8800632Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8800833Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8801031Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8801232Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8801448Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8801767Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8801976Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8802185Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8802416Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8802613Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8802818Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8803023Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8803225Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8803424Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8803630Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8803834Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8804046Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8804245Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8804442Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8804630Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8804827Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8805026Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8805217Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8805414Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8805608Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8805806Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8805998Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8806196Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8806391Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8806584Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8806783Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8807094Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8807291Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8807489Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8807688Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8807877Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8808074Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8808353Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8808553Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8808745Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8808943Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8809140Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8809330Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8809533Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8809735Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8809931Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8810131Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8810328Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8810521Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8810724Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8810925Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8811123Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8811321Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8811516Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8811716Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8811908Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8812105Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8812305Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8812496Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8812695Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8812897Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8813094Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8813297Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8813499Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8813699Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8813902Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8814106Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8814309Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8814502Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8814704Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8814905Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8815099Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8815302Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8815505Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8815702Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8815938Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8816139Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8816340Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8816543Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8816738Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8816940Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8817168Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8817373Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8817579Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8817779Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8817973Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8818177Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8818381Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8818578Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8818782Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8818986Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8819183Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8819427Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8819636Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8819834Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8820042Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8820244Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8820447Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8820677Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8820882Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8821089Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8821284Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8821485Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8821685Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8821878Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8822081Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8822283Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8822479Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8822682Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8822883Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8823084Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8823279Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8823482Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8823684Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8823878Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8824079Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8824283Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8824478Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8824681Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8824884Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8825078Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8825280Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8825485Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8825690Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8825889Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8826092Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8826293Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8826489Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8826696Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8826896Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8840017Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8840279Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8840496Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8840705Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8840907Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8841119Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8841327Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8841602Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8841834Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8842043Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8842265Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8842501Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8842733Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8842967Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8843183Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8843400Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8843621Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8843819Z cudagraph partition due to non gpu ops 2025-12-04T09:39:16.8844068Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:16.8844462Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:16.8844800Z res = mod(**inputs) 2025-12-04T09:39:16.8845358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/electra/modeling_electra.py", line 1554, in forward 2025-12-04T09:39:16.8845839Z lm_loss = self.loss_function( 2025-12-04T09:39:16.8846231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/loss/loss_utils.py", line 67, in ForCausalLMLoss 2025-12-04T09:39:16.8846775Z loss = fixed_cross_entropy(logits, shift_labels, num_items_in_batch, ignore_index, **kwargs) 2025-12-04T09:39:16.8847265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/loss/loss_utils.py", line 36, in fixed_cross_entropy 2025-12-04T09:39:16.8847780Z loss = nn.functional.cross_entropy(source, target, ignore_index=ignore_index, reduction=reduction) 2025-12-04T09:39:16.8848040Z 2025-12-04T09:39:25.8109007Z Compilation time (from dynamo_timed): 17.957642925 2025-12-04T09:39:25.8144346Z pass 2025-12-04T09:39:25.8148813Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:39:25.8149646Z TIMING: _recursive_pre_grad_passes:0.03028 _recursive_joint_graph_passes:0.45012 _recursive_post_grad_passes:0.06527 async_compile.wait:0.78139 code_gen:8.49063 inductor_compile:10.31641 backend_compile:15.09066 gc:0.00022 entire_frame_compile:17.95764 total_wall_time:17.95764 2025-12-04T09:39:25.8150561Z STATS: call_* op count: 379 | FakeTensorMode.__torch_dispatch__:22296 | FakeTensor.__torch_dispatch__:2889 | ProxyTorchDispatchMode.__torch_dispatch__:4842 2025-12-04T09:39:25.8151071Z Dynamo produced 1 graphs covering 379 ops with 0 graph breaks (0 unique) 2025-12-04T09:39:28.1480809Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:39:28.1481887Z import pynvml # type: ignore[import] 2025-12-04T09:39:31.3112264Z 2025-12-04T09:39:32.6022290Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:39:32.6025056Z loading model: 0it [00:01, ?it/s] 2025-12-04T09:39:32.6025583Z cpu eval GPT2ForSequenceClassification 2025-12-04T09:39:33.4312303Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:39:33.6112991Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:39:33.7757965Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:39:41.7657807Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7662115Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7664429Z res = mod(**inputs) 2025-12-04T09:39:41.7667522Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7668097Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7673383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 873, in forward 2025-12-04T09:39:41.7675226Z causal_mask = create_causal_mask( 2025-12-04T09:39:41.7678000Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 788, in create_causal_mask 2025-12-04T09:39:41.7678546Z early_exit, attention_mask, packed_sequence_mask, kv_length, kv_offset = _preprocess_mask_arguments( 2025-12-04T09:39:41.7679074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 740, in _preprocess_mask_arguments 2025-12-04T09:39:41.7679545Z packed_sequence_mask = find_packed_sequence_indices(position_ids) 2025-12-04T09:39:41.7680011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 655, in find_packed_sequence_indices 2025-12-04T09:39:41.7680697Z first_dummy_value = position_ids[:, :1] - 1 # We just need the diff on this first value to be 1 2025-12-04T09:39:41.7680949Z 2025-12-04T09:39:41.7681054Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.7681304Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7681749Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7682196Z res = mod(**inputs) 2025-12-04T09:39:41.7682563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1394, in forward 2025-12-04T09:39:41.7683013Z last_non_pad_token = (token_indices * non_pad_mask).argmax(-1) 2025-12-04T09:39:41.7683199Z 2025-12-04T09:39:41.7683307Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7683670Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7683989Z res = mod(**inputs) 2025-12-04T09:39:41.7684359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7684801Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7685194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7685574Z outputs = block( 2025-12-04T09:39:41.7685910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7686277Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7686659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7687048Z return func(*args, **kwargs) 2025-12-04T09:39:41.7687412Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7687811Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7688200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7688563Z return func(*args, **kwargs) 2025-12-04T09:39:41.7688931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.7694758Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.7695336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7695757Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7695940Z 2025-12-04T09:39:41.7696069Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7696519Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7696860Z res = mod(**inputs) 2025-12-04T09:39:41.7697267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7697704Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7698118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7698555Z outputs = block( 2025-12-04T09:39:41.7698901Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7699287Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7699684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7700074Z return func(*args, **kwargs) 2025-12-04T09:39:41.7700703Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7701150Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7701609Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7702008Z return func(*args, **kwargs) 2025-12-04T09:39:41.7702479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.7702970Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.7703438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7703847Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7704024Z 2025-12-04T09:39:41.7704141Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7704513Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7704843Z res = mod(**inputs) 2025-12-04T09:39:41.7705209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7705620Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7706029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 873, in forward 2025-12-04T09:39:41.7706421Z causal_mask = create_causal_mask( 2025-12-04T09:39:41.7706811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 788, in create_causal_mask 2025-12-04T09:39:41.7707315Z early_exit, attention_mask, packed_sequence_mask, kv_length, kv_offset = _preprocess_mask_arguments( 2025-12-04T09:39:41.7707854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 740, in _preprocess_mask_arguments 2025-12-04T09:39:41.7708334Z packed_sequence_mask = find_packed_sequence_indices(position_ids) 2025-12-04T09:39:41.7708800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 656, in find_packed_sequence_indices 2025-12-04T09:39:41.7709278Z position_diff = torch.diff(position_ids, prepend=first_dummy_value, dim=-1) 2025-12-04T09:39:41.7709509Z 2025-12-04T09:39:41.7709595Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.7709841Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7710204Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7710520Z res = mod(**inputs) 2025-12-04T09:39:41.7710883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7711278Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7711662Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 873, in forward 2025-12-04T09:39:41.7712053Z causal_mask = create_causal_mask( 2025-12-04T09:39:41.7712425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 788, in create_causal_mask 2025-12-04T09:39:41.7712969Z early_exit, attention_mask, packed_sequence_mask, kv_length, kv_offset = _preprocess_mask_arguments( 2025-12-04T09:39:41.7713494Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 740, in _preprocess_mask_arguments 2025-12-04T09:39:41.7713946Z packed_sequence_mask = find_packed_sequence_indices(position_ids) 2025-12-04T09:39:41.7714401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 657, in find_packed_sequence_indices 2025-12-04T09:39:41.7714837Z packed_sequence_mask = (position_diff != 1).cumsum(-1) 2025-12-04T09:39:41.7715003Z 2025-12-04T09:39:41.7715160Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7715524Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7715842Z res = mod(**inputs) 2025-12-04T09:39:41.7716190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7716637Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7717027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 873, in forward 2025-12-04T09:39:41.7717411Z causal_mask = create_causal_mask( 2025-12-04T09:39:41.7717779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 788, in create_causal_mask 2025-12-04T09:39:41.7718291Z early_exit, attention_mask, packed_sequence_mask, kv_length, kv_offset = _preprocess_mask_arguments( 2025-12-04T09:39:41.7718833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 740, in _preprocess_mask_arguments 2025-12-04T09:39:41.7719307Z packed_sequence_mask = find_packed_sequence_indices(position_ids) 2025-12-04T09:39:41.7719799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 657, in find_packed_sequence_indices 2025-12-04T09:39:41.7720247Z packed_sequence_mask = (position_diff != 1).cumsum(-1) 2025-12-04T09:39:41.7720413Z 2025-12-04T09:39:41.7720528Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7720896Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7721229Z res = mod(**inputs) 2025-12-04T09:39:41.7721785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7722234Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7722664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7723071Z outputs = block( 2025-12-04T09:39:41.7723434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7723829Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7724214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7724603Z return func(*args, **kwargs) 2025-12-04T09:39:41.7724985Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7725396Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7725785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7726163Z return func(*args, **kwargs) 2025-12-04T09:39:41.7726531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:41.7726929Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:41.7727376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:41.7727865Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:41.7728049Z 2025-12-04T09:39:41.7728139Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.7728375Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7728730Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7729046Z res = mod(**inputs) 2025-12-04T09:39:41.7729434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7729827Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7730469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7730835Z outputs = block( 2025-12-04T09:39:41.7731145Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7731601Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7731972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7732324Z return func(*args, **kwargs) 2025-12-04T09:39:41.7732683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7733067Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7733449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7733806Z return func(*args, **kwargs) 2025-12-04T09:39:41.7734161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:41.7734554Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:41.7734994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:41.7735450Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:41.7735633Z 2025-12-04T09:39:41.7735736Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7736090Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7736396Z res = mod(**inputs) 2025-12-04T09:39:41.7736745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7737127Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7737502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7737852Z outputs = block( 2025-12-04T09:39:41.7738168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7738524Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7738888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7739257Z return func(*args, **kwargs) 2025-12-04T09:39:41.7739624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7740020Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7740398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7740769Z return func(*args, **kwargs) 2025-12-04T09:39:41.7741137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.7741526Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.7741879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7742280Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7742453Z 2025-12-04T09:39:41.7742568Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7742926Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7743254Z res = mod(**inputs) 2025-12-04T09:39:41.7743658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7744047Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7744419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7744779Z outputs = block( 2025-12-04T09:39:41.7745180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7745562Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7745935Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7746298Z return func(*args, **kwargs) 2025-12-04T09:39:41.7746692Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7747079Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7747457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7747813Z return func(*args, **kwargs) 2025-12-04T09:39:41.7748174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.7748555Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.7748898Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7749289Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7749465Z 2025-12-04T09:39:41.7749567Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7749922Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7750231Z res = mod(**inputs) 2025-12-04T09:39:41.7750583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7750968Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7751350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7751708Z outputs = block( 2025-12-04T09:39:41.7752023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7752377Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7752749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7753124Z return func(*args, **kwargs) 2025-12-04T09:39:41.7753490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7753916Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7754311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.7754692Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.7755048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7755437Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7755617Z 2025-12-04T09:39:41.7755722Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7756084Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7756403Z res = mod(**inputs) 2025-12-04T09:39:41.7756753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7757186Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7757577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7757946Z outputs = block( 2025-12-04T09:39:41.7758261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7758660Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7759039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7759404Z return func(*args, **kwargs) 2025-12-04T09:39:41.7759778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7760187Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7760600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.7760982Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.7761339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7761807Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7761997Z 2025-12-04T09:39:41.7762121Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7762506Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7762871Z res = mod(**inputs) 2025-12-04T09:39:41.7763263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7763685Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7764106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7764480Z outputs = block( 2025-12-04T09:39:41.7764807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7765164Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7765546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7765926Z return func(*args, **kwargs) 2025-12-04T09:39:41.7766289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7766709Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7767118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:41.7767509Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:41.7767858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:41.7768323Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:41.7768556Z 2025-12-04T09:39:41.7768667Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7769038Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7769349Z res = mod(**inputs) 2025-12-04T09:39:41.7769700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7770089Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7770458Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7770819Z outputs = block( 2025-12-04T09:39:41.7771174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7771527Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7771888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7772251Z return func(*args, **kwargs) 2025-12-04T09:39:41.7772643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7773039Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7773436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.7773820Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.7774180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7774566Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7774745Z 2025-12-04T09:39:41.7774848Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7775199Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7775511Z res = mod(**inputs) 2025-12-04T09:39:41.7775851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7776237Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7776613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7776963Z outputs = block( 2025-12-04T09:39:41.7777637Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7777995Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7778367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7778720Z return func(*args, **kwargs) 2025-12-04T09:39:41.7779078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7779495Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7779886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.7780269Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.7780623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7781008Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7781175Z 2025-12-04T09:39:41.7781277Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7781637Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7781942Z res = mod(**inputs) 2025-12-04T09:39:41.7782274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7782646Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7783018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7783375Z outputs = block( 2025-12-04T09:39:41.7783684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7784041Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7784412Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7784768Z return func(*args, **kwargs) 2025-12-04T09:39:41.7785164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7785553Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7785919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7786307Z return func(*args, **kwargs) 2025-12-04T09:39:41.7786664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.7787145Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.7787601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7787966Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7788133Z 2025-12-04T09:39:41.7788237Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7788585Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7788895Z res = mod(**inputs) 2025-12-04T09:39:41.7789232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7789610Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7789997Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7790336Z outputs = block( 2025-12-04T09:39:41.7790647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7790997Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7791361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7791718Z return func(*args, **kwargs) 2025-12-04T09:39:41.7792076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7792458Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7792823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7793185Z return func(*args, **kwargs) 2025-12-04T09:39:41.7793542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.7794017Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.7794452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7794849Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7795009Z 2025-12-04T09:39:41.7795096Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.7795319Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7795662Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7795987Z res = mod(**inputs) 2025-12-04T09:39:41.7796335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7796709Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7797082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7797439Z outputs = block( 2025-12-04T09:39:41.7797745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7798145Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7798519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7798882Z return func(*args, **kwargs) 2025-12-04T09:39:41.7799232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7799650Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7800026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7800382Z return func(*args, **kwargs) 2025-12-04T09:39:41.7800732Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:41.7801123Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:41.7801675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:41.7802205Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:41.7802411Z 2025-12-04T09:39:41.7802532Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7802901Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7803233Z res = mod(**inputs) 2025-12-04T09:39:41.7803589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7803981Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7804368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7804735Z outputs = block( 2025-12-04T09:39:41.7805048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7805412Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7805791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7806152Z return func(*args, **kwargs) 2025-12-04T09:39:41.7806518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7806912Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7807288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7807647Z return func(*args, **kwargs) 2025-12-04T09:39:41.7808008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.7808394Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.7808743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7809135Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7809309Z 2025-12-04T09:39:41.7809413Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7809766Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7810076Z res = mod(**inputs) 2025-12-04T09:39:41.7810433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7810824Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7811203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7811560Z outputs = block( 2025-12-04T09:39:41.7812913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7813277Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7813640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7814007Z return func(*args, **kwargs) 2025-12-04T09:39:41.7814381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7814795Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7815156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7815511Z return func(*args, **kwargs) 2025-12-04T09:39:41.7815860Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.7816221Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.7816562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7816939Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7817099Z 2025-12-04T09:39:41.7817204Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7817539Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7817843Z res = mod(**inputs) 2025-12-04T09:39:41.7818182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7818556Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7818916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7819266Z outputs = block( 2025-12-04T09:39:41.7819573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7819909Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7820272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7820620Z return func(*args, **kwargs) 2025-12-04T09:39:41.7820970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7821350Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7821728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.7822097Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.7822425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7822803Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7822969Z 2025-12-04T09:39:41.7823068Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7823406Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7823698Z res = mod(**inputs) 2025-12-04T09:39:41.7824033Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7824408Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7824773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7825117Z outputs = block( 2025-12-04T09:39:41.7825422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7825767Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7826157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7826517Z return func(*args, **kwargs) 2025-12-04T09:39:41.7826875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7827263Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7827686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.7828066Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.7828412Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7828786Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7828958Z 2025-12-04T09:39:41.7829059Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7829410Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7829715Z res = mod(**inputs) 2025-12-04T09:39:41.7830052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7830561Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7830940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7831295Z outputs = block( 2025-12-04T09:39:41.7831594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7831943Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7832304Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7832651Z return func(*args, **kwargs) 2025-12-04T09:39:41.7833006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7833399Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7833787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:41.7834149Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:41.7834488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:41.7834925Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:41.7835143Z 2025-12-04T09:39:41.7835250Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7835586Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7835894Z res = mod(**inputs) 2025-12-04T09:39:41.7836234Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7836599Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7836970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7837324Z outputs = block( 2025-12-04T09:39:41.7837631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7837969Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7838327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7838680Z return func(*args, **kwargs) 2025-12-04T09:39:41.7839096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7839488Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7839882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.7840256Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.7840592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7841016Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7841184Z 2025-12-04T09:39:41.7841291Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7841710Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7842045Z res = mod(**inputs) 2025-12-04T09:39:41.7842447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7842840Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7843228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7843578Z outputs = block( 2025-12-04T09:39:41.7843883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7844229Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7844582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7844934Z return func(*args, **kwargs) 2025-12-04T09:39:41.7845288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7845668Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7846052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.7846424Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.7846763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7847129Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7847302Z 2025-12-04T09:39:41.7847400Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7847746Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7848046Z res = mod(**inputs) 2025-12-04T09:39:41.7848378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7848747Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7849114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7849455Z outputs = block( 2025-12-04T09:39:41.7849758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7850098Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7850456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7850805Z return func(*args, **kwargs) 2025-12-04T09:39:41.7851150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7851528Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7851892Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7852245Z return func(*args, **kwargs) 2025-12-04T09:39:41.7852643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.7853107Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.7853530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7853941Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7854110Z 2025-12-04T09:39:41.7854208Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7854552Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7854847Z res = mod(**inputs) 2025-12-04T09:39:41.7855189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7855565Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7855933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7856286Z outputs = block( 2025-12-04T09:39:41.7856590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7856932Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7857286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7857643Z return func(*args, **kwargs) 2025-12-04T09:39:41.7857992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7858365Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7858723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7859079Z return func(*args, **kwargs) 2025-12-04T09:39:41.7859429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.7859879Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.7860310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7860698Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7860859Z 2025-12-04T09:39:41.7860944Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.7861168Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7861511Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7861814Z res = mod(**inputs) 2025-12-04T09:39:41.7862149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7862520Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7862885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7863231Z outputs = block( 2025-12-04T09:39:41.7863528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7863871Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7864229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7864578Z return func(*args, **kwargs) 2025-12-04T09:39:41.7864918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7865291Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7865696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7866043Z return func(*args, **kwargs) 2025-12-04T09:39:41.7866392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:41.7866773Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:41.7867230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:41.7867676Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:41.7867857Z 2025-12-04T09:39:41.7867955Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7868297Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7868600Z res = mod(**inputs) 2025-12-04T09:39:41.7868936Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7869310Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7869675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7870016Z outputs = block( 2025-12-04T09:39:41.7870324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7870664Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7871020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7871360Z return func(*args, **kwargs) 2025-12-04T09:39:41.7871705Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7872080Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7872437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7872797Z return func(*args, **kwargs) 2025-12-04T09:39:41.7873144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.7873511Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.7873841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7874219Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7874379Z 2025-12-04T09:39:41.7874484Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7874824Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7875119Z res = mod(**inputs) 2025-12-04T09:39:41.7875457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7875830Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7876189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7876540Z outputs = block( 2025-12-04T09:39:41.7876847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7877190Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7877541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7877891Z return func(*args, **kwargs) 2025-12-04T09:39:41.7878243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7878645Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7879016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7879366Z return func(*args, **kwargs) 2025-12-04T09:39:41.7879715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.7880108Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.7880455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7880839Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7881002Z 2025-12-04T09:39:41.7881110Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7881452Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7881833Z res = mod(**inputs) 2025-12-04T09:39:41.7882196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7882584Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7882989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7883387Z outputs = block( 2025-12-04T09:39:41.7883720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7884081Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7884453Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7884814Z return func(*args, **kwargs) 2025-12-04T09:39:41.7885166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7885573Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7885971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.7886350Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.7886689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7887078Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7887243Z 2025-12-04T09:39:41.7887352Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7887698Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7888003Z res = mod(**inputs) 2025-12-04T09:39:41.7888349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7888734Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7889104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7889463Z outputs = block( 2025-12-04T09:39:41.7889777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7890131Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7890490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7890850Z return func(*args, **kwargs) 2025-12-04T09:39:41.7891207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7891600Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7892036Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.7892422Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.7892766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7893142Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7893361Z 2025-12-04T09:39:41.7893464Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7893812Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7894130Z res = mod(**inputs) 2025-12-04T09:39:41.7894470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7894853Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7895238Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7895581Z outputs = block( 2025-12-04T09:39:41.7895888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7896227Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7896585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7896934Z return func(*args, **kwargs) 2025-12-04T09:39:41.7897285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7897677Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7898098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:41.7898463Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:41.7898795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:41.7899220Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:41.7899437Z 2025-12-04T09:39:41.7899536Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7899876Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7900178Z res = mod(**inputs) 2025-12-04T09:39:41.7900511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7900875Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7901236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7901585Z outputs = block( 2025-12-04T09:39:41.7901883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7902222Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7902578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7902929Z return func(*args, **kwargs) 2025-12-04T09:39:41.7903271Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7903658Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7904042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.7904409Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.7904743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7905147Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7905311Z 2025-12-04T09:39:41.7905417Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7905748Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7906050Z res = mod(**inputs) 2025-12-04T09:39:41.7906423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7906794Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7907154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7907501Z outputs = block( 2025-12-04T09:39:41.7907815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7908162Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7908536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7908885Z return func(*args, **kwargs) 2025-12-04T09:39:41.7909236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7909622Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7910007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.7910382Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.7910726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7911096Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7911265Z 2025-12-04T09:39:41.7911365Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7911706Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7912003Z res = mod(**inputs) 2025-12-04T09:39:41.7912345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7912719Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7913087Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7913429Z outputs = block( 2025-12-04T09:39:41.7913743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7914088Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7914438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7914794Z return func(*args, **kwargs) 2025-12-04T09:39:41.7915143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 451, in forward 2025-12-04T09:39:41.7915534Z hidden_states = residual + feed_forward_hidden_states 2025-12-04T09:39:41.7915685Z 2025-12-04T09:39:41.7915783Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7916125Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7916427Z res = mod(**inputs) 2025-12-04T09:39:41.7916763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7917128Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7917492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7917847Z outputs = block( 2025-12-04T09:39:41.7918194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7918541Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7918899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7919248Z return func(*args, **kwargs) 2025-12-04T09:39:41.7919625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7919995Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7920364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7920707Z return func(*args, **kwargs) 2025-12-04T09:39:41.7921053Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.7921523Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.7922072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7922487Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7922667Z 2025-12-04T09:39:41.7922774Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7923143Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7923457Z res = mod(**inputs) 2025-12-04T09:39:41.7923807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7924183Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7924562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7924918Z outputs = block( 2025-12-04T09:39:41.7925237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7925597Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7925966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7926329Z return func(*args, **kwargs) 2025-12-04T09:39:41.7926689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7927074Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7927451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7927807Z return func(*args, **kwargs) 2025-12-04T09:39:41.7928169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.7928648Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.7929097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7929491Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7929666Z 2025-12-04T09:39:41.7929748Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.7929988Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7930503Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7930823Z res = mod(**inputs) 2025-12-04T09:39:41.7931176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7931560Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7932000Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7932362Z outputs = block( 2025-12-04T09:39:41.7932676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7933017Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7933437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7933803Z return func(*args, **kwargs) 2025-12-04T09:39:41.7934167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7934560Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7934943Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7935315Z return func(*args, **kwargs) 2025-12-04T09:39:41.7935669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:41.7936069Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:41.7936508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:41.7936979Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:41.7937155Z 2025-12-04T09:39:41.7937257Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7937621Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7937932Z res = mod(**inputs) 2025-12-04T09:39:41.7938276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7938649Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7939018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7939370Z outputs = block( 2025-12-04T09:39:41.7939672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7940017Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7940381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7940736Z return func(*args, **kwargs) 2025-12-04T09:39:41.7941082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7941465Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7941840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7942187Z return func(*args, **kwargs) 2025-12-04T09:39:41.7942542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.7942914Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.7943257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7943635Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7943805Z 2025-12-04T09:39:41.7943907Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7944257Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7944566Z res = mod(**inputs) 2025-12-04T09:39:41.7944905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7945319Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7945695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7946041Z outputs = block( 2025-12-04T09:39:41.7946352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7946741Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7947100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7947442Z return func(*args, **kwargs) 2025-12-04T09:39:41.7947788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7948158Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7948517Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7948864Z return func(*args, **kwargs) 2025-12-04T09:39:41.7949208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.7949575Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.7949908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7950286Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7950450Z 2025-12-04T09:39:41.7950556Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7950898Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7951192Z res = mod(**inputs) 2025-12-04T09:39:41.7951532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7951903Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7952261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7952608Z outputs = block( 2025-12-04T09:39:41.7952914Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7953259Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7953610Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7953957Z return func(*args, **kwargs) 2025-12-04T09:39:41.7954302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7954684Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7955071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.7955435Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.7955767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7956131Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7956299Z 2025-12-04T09:39:41.7956397Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7956733Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7957036Z res = mod(**inputs) 2025-12-04T09:39:41.7957369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7957740Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7958141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7958490Z outputs = block( 2025-12-04T09:39:41.7958802Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7959154Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7959543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7959885Z return func(*args, **kwargs) 2025-12-04T09:39:41.7960232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7960620Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7960993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.7961363Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.7961763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7962154Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7962318Z 2025-12-04T09:39:41.7962419Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7962773Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7963088Z res = mod(**inputs) 2025-12-04T09:39:41.7963444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7963834Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7964217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7964600Z outputs = block( 2025-12-04T09:39:41.7964924Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7965289Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7965667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7966039Z return func(*args, **kwargs) 2025-12-04T09:39:41.7966401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7966813Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7967218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:41.7967600Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:41.7967951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:41.7968405Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:41.7968634Z 2025-12-04T09:39:41.7968746Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7969102Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7969429Z res = mod(**inputs) 2025-12-04T09:39:41.7969789Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7970180Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7970558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7970927Z outputs = block( 2025-12-04T09:39:41.7971249Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7971637Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7972024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7972374Z return func(*args, **kwargs) 2025-12-04T09:39:41.7972721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7973131Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7973515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.7973888Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.7974228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7974596Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7974763Z 2025-12-04T09:39:41.7974867Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7975208Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7975502Z res = mod(**inputs) 2025-12-04T09:39:41.7975840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7976215Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7976580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7976919Z outputs = block( 2025-12-04T09:39:41.7977224Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7977568Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7977919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7978269Z return func(*args, **kwargs) 2025-12-04T09:39:41.7978616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.7979003Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.7979384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.7979768Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.7980117Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7980500Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7980664Z 2025-12-04T09:39:41.7980764Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7981135Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7981440Z res = mod(**inputs) 2025-12-04T09:39:41.7981770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7982140Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7982508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7982859Z outputs = block( 2025-12-04T09:39:41.7983155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7983499Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7983856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7984198Z return func(*args, **kwargs) 2025-12-04T09:39:41.7984587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7984966Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7985335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7985681Z return func(*args, **kwargs) 2025-12-04T09:39:41.7986031Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.7986532Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.7986964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7987335Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7987505Z 2025-12-04T09:39:41.7987605Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7987954Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7988252Z res = mod(**inputs) 2025-12-04T09:39:41.7988592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7988972Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7989352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7989701Z outputs = block( 2025-12-04T09:39:41.7990014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7990365Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7990736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7991077Z return func(*args, **kwargs) 2025-12-04T09:39:41.7991488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7991872Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7992241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7992606Z return func(*args, **kwargs) 2025-12-04T09:39:41.7992964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.7993435Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.7993870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.7994259Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.7994423Z 2025-12-04T09:39:41.7994514Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.7994749Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.7995098Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.7995410Z res = mod(**inputs) 2025-12-04T09:39:41.7995761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.7996143Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.7996526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.7996883Z outputs = block( 2025-12-04T09:39:41.7997196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.7997544Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.7997963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7998326Z return func(*args, **kwargs) 2025-12-04T09:39:41.7998678Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.7999059Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.7999473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.7999834Z return func(*args, **kwargs) 2025-12-04T09:39:41.8000184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:41.8000582Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:41.8001019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:41.8001494Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:41.8001737Z 2025-12-04T09:39:41.8001847Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8002223Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8002556Z res = mod(**inputs) 2025-12-04T09:39:41.8002922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8003336Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8003741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8004127Z outputs = block( 2025-12-04T09:39:41.8004458Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8004816Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8005193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8005549Z return func(*args, **kwargs) 2025-12-04T09:39:41.8005907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8006292Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8006664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8007009Z return func(*args, **kwargs) 2025-12-04T09:39:41.8007357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8007724Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8008064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8008441Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8008611Z 2025-12-04T09:39:41.8008710Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8009054Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8009352Z res = mod(**inputs) 2025-12-04T09:39:41.8009699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8010076Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8010447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8010791Z outputs = block( 2025-12-04T09:39:41.8011097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8011445Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8011834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8012185Z return func(*args, **kwargs) 2025-12-04T09:39:41.8012535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8012941Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8013299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8013648Z return func(*args, **kwargs) 2025-12-04T09:39:41.8013991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8014356Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8014689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8015069Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8015230Z 2025-12-04T09:39:41.8015339Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8015669Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8015970Z res = mod(**inputs) 2025-12-04T09:39:41.8016316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8016685Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8017044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8017393Z outputs = block( 2025-12-04T09:39:41.8017695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8018033Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8018388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8018734Z return func(*args, **kwargs) 2025-12-04T09:39:41.8019078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8019462Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8019845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8020211Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8020546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8020912Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8021079Z 2025-12-04T09:39:41.8021178Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8021514Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8021810Z res = mod(**inputs) 2025-12-04T09:39:41.8022146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8022519Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8022882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8023221Z outputs = block( 2025-12-04T09:39:41.8023524Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8023865Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8024214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8024599Z return func(*args, **kwargs) 2025-12-04T09:39:41.8024952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8025342Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8025722Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8026120Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8026456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8026829Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8026994Z 2025-12-04T09:39:41.8027092Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8027431Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8027748Z res = mod(**inputs) 2025-12-04T09:39:41.8028089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8028470Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8028835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8029188Z outputs = block( 2025-12-04T09:39:41.8029487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8029826Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8030365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8030719Z return func(*args, **kwargs) 2025-12-04T09:39:41.8031076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8031468Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8031855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:41.8032218Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:41.8032554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:41.8032988Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:41.8033206Z 2025-12-04T09:39:41.8033313Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8033649Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8033957Z res = mod(**inputs) 2025-12-04T09:39:41.8034299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8034666Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8035032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8035381Z outputs = block( 2025-12-04T09:39:41.8035687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8036030Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8036389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8036742Z return func(*args, **kwargs) 2025-12-04T09:39:41.8037091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8037472Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8037925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8038296Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8038630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8039059Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8039230Z 2025-12-04T09:39:41.8039333Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8039685Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8039989Z res = mod(**inputs) 2025-12-04T09:39:41.8040369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8040749Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8041129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8041513Z outputs = block( 2025-12-04T09:39:41.8041900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8042274Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8042656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8043016Z return func(*args, **kwargs) 2025-12-04T09:39:41.8043376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8043774Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8044167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8044557Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8044909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8045290Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8045466Z 2025-12-04T09:39:41.8045568Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8045921Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8046237Z res = mod(**inputs) 2025-12-04T09:39:41.8046577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8046958Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8047333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8047688Z outputs = block( 2025-12-04T09:39:41.8048011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8048363Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8048731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8049087Z return func(*args, **kwargs) 2025-12-04T09:39:41.8049442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 451, in forward 2025-12-04T09:39:41.8049847Z hidden_states = residual + feed_forward_hidden_states 2025-12-04T09:39:41.8050003Z 2025-12-04T09:39:41.8050112Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8050448Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8050759Z res = mod(**inputs) 2025-12-04T09:39:41.8051160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8051535Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8051913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8052266Z outputs = block( 2025-12-04T09:39:41.8052615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8052969Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8053331Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8053691Z return func(*args, **kwargs) 2025-12-04T09:39:41.8054032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8054409Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8054782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8055131Z return func(*args, **kwargs) 2025-12-04T09:39:41.8055473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.8055944Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.8056382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8056760Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8056924Z 2025-12-04T09:39:41.8057023Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8057362Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8057665Z res = mod(**inputs) 2025-12-04T09:39:41.8057906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8057995Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8058226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8058289Z outputs = block( 2025-12-04T09:39:41.8058503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8058581Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8058810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8058885Z return func(*args, **kwargs) 2025-12-04T09:39:41.8059118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8059211Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8059440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8059507Z return func(*args, **kwargs) 2025-12-04T09:39:41.8059743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.8059917Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.8060130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8060244Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8060247Z 2025-12-04T09:39:41.8060326Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.8060434Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8060655Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8060718Z res = mod(**inputs) 2025-12-04T09:39:41.8060963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8061042Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8061317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8061379Z outputs = block( 2025-12-04T09:39:41.8061591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8061678Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8061911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8061981Z return func(*args, **kwargs) 2025-12-04T09:39:41.8062236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8062321Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8062561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8062630Z return func(*args, **kwargs) 2025-12-04T09:39:41.8062858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:41.8062959Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:41.8063230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:41.8063362Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:41.8063365Z 2025-12-04T09:39:41.8063470Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8063658Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8063728Z res = mod(**inputs) 2025-12-04T09:39:41.8063968Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8064052Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8064296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8064359Z outputs = block( 2025-12-04T09:39:41.8064585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8064663Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8064896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8064975Z return func(*args, **kwargs) 2025-12-04T09:39:41.8065217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8065306Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8065528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8065597Z return func(*args, **kwargs) 2025-12-04T09:39:41.8065834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8065910Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8066115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8066233Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8066237Z 2025-12-04T09:39:41.8066367Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8066559Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8066621Z res = mod(**inputs) 2025-12-04T09:39:41.8066855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8066974Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8067205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8067265Z outputs = block( 2025-12-04T09:39:41.8067484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8067562Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8067808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8067876Z return func(*args, **kwargs) 2025-12-04T09:39:41.8068124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8068215Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8068442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8068515Z return func(*args, **kwargs) 2025-12-04T09:39:41.8068752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8068831Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8069047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8069160Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8069163Z 2025-12-04T09:39:41.8069275Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8069465Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8069525Z res = mod(**inputs) 2025-12-04T09:39:41.8069766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8069846Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8070078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8070143Z outputs = block( 2025-12-04T09:39:41.8070353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8070426Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8070659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8070723Z return func(*args, **kwargs) 2025-12-04T09:39:41.8070960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8071057Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8071290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8071372Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8071577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8071694Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8071697Z 2025-12-04T09:39:41.8071793Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8072008Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8072077Z res = mod(**inputs) 2025-12-04T09:39:41.8072312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8072390Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8072623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8072724Z outputs = block( 2025-12-04T09:39:41.8072937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8073012Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8073235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8073307Z return func(*args, **kwargs) 2025-12-04T09:39:41.8073538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8073645Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8073875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8073950Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8074161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8074271Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8074274Z 2025-12-04T09:39:41.8074371Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8074563Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8074624Z res = mod(**inputs) 2025-12-04T09:39:41.8074865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8074944Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8075176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8075244Z outputs = block( 2025-12-04T09:39:41.8075450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8075534Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8075759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8075826Z return func(*args, **kwargs) 2025-12-04T09:39:41.8076062Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8076159Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8076388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:41.8076471Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:41.8076669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:41.8076850Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:41.8076853Z 2025-12-04T09:39:41.8076954Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8077138Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8077208Z res = mod(**inputs) 2025-12-04T09:39:41.8077448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8077533Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8077799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8077862Z outputs = block( 2025-12-04T09:39:41.8078083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8078192Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8078427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8078501Z return func(*args, **kwargs) 2025-12-04T09:39:41.8078737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8078843Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8079085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8079170Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8079385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8079497Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8079500Z 2025-12-04T09:39:41.8079613Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8079806Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8079869Z res = mod(**inputs) 2025-12-04T09:39:41.8080116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8080196Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8080429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8080499Z outputs = block( 2025-12-04T09:39:41.8080710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8080794Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8081025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8081097Z return func(*args, **kwargs) 2025-12-04T09:39:41.8081339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8081437Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8081739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8081845Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8082086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8082222Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8082226Z 2025-12-04T09:39:41.8082338Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8082552Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8082629Z res = mod(**inputs) 2025-12-04T09:39:41.8082885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8082988Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8083225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8083286Z outputs = block( 2025-12-04T09:39:41.8083507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8083626Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8083859Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8083936Z return func(*args, **kwargs) 2025-12-04T09:39:41.8084172Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8084299Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8084528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8084595Z return func(*args, **kwargs) 2025-12-04T09:39:41.8084839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.8085015Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.8085234Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8085348Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8085351Z 2025-12-04T09:39:41.8085451Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8085651Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8085712Z res = mod(**inputs) 2025-12-04T09:39:41.8085962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8086047Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8086284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8086353Z outputs = block( 2025-12-04T09:39:41.8086569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8086645Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8086885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8086951Z return func(*args, **kwargs) 2025-12-04T09:39:41.8087190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8087284Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8087517Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8087589Z return func(*args, **kwargs) 2025-12-04T09:39:41.8087825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.8088005Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.8088222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8088335Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8088338Z 2025-12-04T09:39:41.8088429Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.8088532Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8088723Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8088791Z res = mod(**inputs) 2025-12-04T09:39:41.8089034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8089112Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8089398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8089462Z outputs = block( 2025-12-04T09:39:41.8089687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8089764Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8090000Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8090105Z return func(*args, **kwargs) 2025-12-04T09:39:41.8090347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8090440Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8090680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8090748Z return func(*args, **kwargs) 2025-12-04T09:39:41.8091002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:41.8091099Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:41.8091388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:41.8091525Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:41.8091529Z 2025-12-04T09:39:41.8091630Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8091834Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8091895Z res = mod(**inputs) 2025-12-04T09:39:41.8092147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8092234Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8092481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8092549Z outputs = block( 2025-12-04T09:39:41.8092769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8092843Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8093090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8093156Z return func(*args, **kwargs) 2025-12-04T09:39:41.8093397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8093490Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8093726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8093798Z return func(*args, **kwargs) 2025-12-04T09:39:41.8094045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8094124Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8094343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8094458Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8094461Z 2025-12-04T09:39:41.8094570Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8094769Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8094832Z res = mod(**inputs) 2025-12-04T09:39:41.8095084Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8095165Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8095440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8095512Z outputs = block( 2025-12-04T09:39:41.8095723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8095807Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8096082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8096149Z return func(*args, **kwargs) 2025-12-04T09:39:41.8096401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8096482Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8096706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8096783Z return func(*args, **kwargs) 2025-12-04T09:39:41.8097014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8097098Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8097301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8097413Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8097416Z 2025-12-04T09:39:41.8097521Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8097707Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8097773Z res = mod(**inputs) 2025-12-04T09:39:41.8098012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8098090Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8098337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8098398Z outputs = block( 2025-12-04T09:39:41.8098610Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8098693Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8098926Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8099000Z return func(*args, **kwargs) 2025-12-04T09:39:41.8099243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8099340Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8099574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8099651Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8099856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8099971Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8099974Z 2025-12-04T09:39:41.8100071Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8100262Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8100324Z res = mod(**inputs) 2025-12-04T09:39:41.8100556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8100642Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8100874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8100973Z outputs = block( 2025-12-04T09:39:41.8101184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8101257Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8101488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8101586Z return func(*args, **kwargs) 2025-12-04T09:39:41.8101815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8101920Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8102150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8102234Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8102440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8102548Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8102551Z 2025-12-04T09:39:41.8102657Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8102838Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8102909Z res = mod(**inputs) 2025-12-04T09:39:41.8103143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8103221Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8103455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8103515Z outputs = block( 2025-12-04T09:39:41.8103720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8103806Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8104028Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8104102Z return func(*args, **kwargs) 2025-12-04T09:39:41.8104332Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8104431Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8104666Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:41.8104744Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:41.8104943Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:41.8105122Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:41.8105128Z 2025-12-04T09:39:41.8105226Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8105416Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8105478Z res = mod(**inputs) 2025-12-04T09:39:41.8105709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8105798Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8106027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8106095Z outputs = block( 2025-12-04T09:39:41.8106298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8106380Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8106641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8106709Z return func(*args, **kwargs) 2025-12-04T09:39:41.8106938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8107043Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8107300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8107388Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8107588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8107698Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8107701Z 2025-12-04T09:39:41.8107807Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8107991Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8108061Z res = mod(**inputs) 2025-12-04T09:39:41.8108293Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8108370Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8108606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8108670Z outputs = block( 2025-12-04T09:39:41.8108874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8108957Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8109183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8109256Z return func(*args, **kwargs) 2025-12-04T09:39:41.8109488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8109586Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8109823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8109905Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8110117Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8110226Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8110229Z 2025-12-04T09:39:41.8110327Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8110517Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8110579Z res = mod(**inputs) 2025-12-04T09:39:41.8110815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8110900Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8111127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8111193Z outputs = block( 2025-12-04T09:39:41.8111402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8111474Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8111704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8111769Z return func(*args, **kwargs) 2025-12-04T09:39:41.8112008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 451, in forward 2025-12-04T09:39:41.8112144Z hidden_states = residual + feed_forward_hidden_states 2025-12-04T09:39:41.8112148Z 2025-12-04T09:39:41.8112246Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8112439Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8112499Z res = mod(**inputs) 2025-12-04T09:39:41.8112729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8112848Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8113076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8113140Z outputs = block( 2025-12-04T09:39:41.8113343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8113416Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8113649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8113715Z return func(*args, **kwargs) 2025-12-04T09:39:41.8113946Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8114037Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8114270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8114343Z return func(*args, **kwargs) 2025-12-04T09:39:41.8114578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.8114755Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.8114975Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8115088Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8115091Z 2025-12-04T09:39:41.8115200Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8115390Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8115453Z res = mod(**inputs) 2025-12-04T09:39:41.8115703Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8115784Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8116026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8116093Z outputs = block( 2025-12-04T09:39:41.8116299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8116385Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8116618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8116684Z return func(*args, **kwargs) 2025-12-04T09:39:41.8116927Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8117014Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8117253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8117320Z return func(*args, **kwargs) 2025-12-04T09:39:41.8117559Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.8117740Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.8117991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8118107Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8118118Z 2025-12-04T09:39:41.8118200Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.8118301Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8118528Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8118592Z res = mod(**inputs) 2025-12-04T09:39:41.8118833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8118921Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8119161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8119229Z outputs = block( 2025-12-04T09:39:41.8119445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8119522Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8119762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8119828Z return func(*args, **kwargs) 2025-12-04T09:39:41.8120069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8120160Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8120392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8120465Z return func(*args, **kwargs) 2025-12-04T09:39:41.8120704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:41.8120802Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:41.8121093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:41.8121221Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:41.8121225Z 2025-12-04T09:39:41.8121332Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8121521Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8121663Z res = mod(**inputs) 2025-12-04T09:39:41.8121921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8122003Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8122241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8122315Z outputs = block( 2025-12-04T09:39:41.8122532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8122617Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8122847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8122916Z return func(*args, **kwargs) 2025-12-04T09:39:41.8123180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8123272Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8123569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8123646Z return func(*args, **kwargs) 2025-12-04T09:39:41.8123928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8124021Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8124237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8124353Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8124356Z 2025-12-04T09:39:41.8124505Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8124701Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8124772Z res = mod(**inputs) 2025-12-04T09:39:41.8125029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8125109Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8125409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8125473Z outputs = block( 2025-12-04T09:39:41.8125682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8125767Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8125998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8126075Z return func(*args, **kwargs) 2025-12-04T09:39:41.8126311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8126394Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8126629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8126695Z return func(*args, **kwargs) 2025-12-04T09:39:41.8126933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8127018Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8127230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8127352Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8127358Z 2025-12-04T09:39:41.8127458Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8127649Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8127718Z res = mod(**inputs) 2025-12-04T09:39:41.8127960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8128046Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8128286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8128345Z outputs = block( 2025-12-04T09:39:41.8128564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8128639Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8128870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8128946Z return func(*args, **kwargs) 2025-12-04T09:39:41.8129185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8129292Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8129528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8129606Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8129855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8129969Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8129972Z 2025-12-04T09:39:41.8130081Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8130443Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8130576Z res = mod(**inputs) 2025-12-04T09:39:41.8130826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8130907Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8131150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8131219Z outputs = block( 2025-12-04T09:39:41.8131429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8131511Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8131737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8131803Z return func(*args, **kwargs) 2025-12-04T09:39:41.8132040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8132142Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8132375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8132460Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8132663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8132783Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8132786Z 2025-12-04T09:39:41.8132884Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8133069Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8133139Z res = mod(**inputs) 2025-12-04T09:39:41.8133374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8133464Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8133694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8133754Z outputs = block( 2025-12-04T09:39:41.8133966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8134041Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8134267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8134340Z return func(*args, **kwargs) 2025-12-04T09:39:41.8134567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8134671Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8134904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:41.8134980Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:41.8135189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:41.8135357Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:41.8135360Z 2025-12-04T09:39:41.8135464Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8135693Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8135755Z res = mod(**inputs) 2025-12-04T09:39:41.8135994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8136070Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8136332Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8136400Z outputs = block( 2025-12-04T09:39:41.8136605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8136687Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8136911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8136980Z return func(*args, **kwargs) 2025-12-04T09:39:41.8137222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8137317Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8137553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8137636Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8137841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8137957Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8137960Z 2025-12-04T09:39:41.8138058Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8138246Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8138316Z res = mod(**inputs) 2025-12-04T09:39:41.8138554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8138640Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8138872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8138934Z outputs = block( 2025-12-04T09:39:41.8139149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8139223Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8139457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8139523Z return func(*args, **kwargs) 2025-12-04T09:39:41.8139753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8139859Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8140091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8140171Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8140383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8140498Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8140501Z 2025-12-04T09:39:41.8140606Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8140790Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8140851Z res = mod(**inputs) 2025-12-04T09:39:41.8141094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8141222Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8141454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8141523Z outputs = block( 2025-12-04T09:39:41.8141730Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8141845Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8142070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8142136Z return func(*args, **kwargs) 2025-12-04T09:39:41.8142380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8142464Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8142696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8142761Z return func(*args, **kwargs) 2025-12-04T09:39:41.8142991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.8143172Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.8143378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8143487Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8143497Z 2025-12-04T09:39:41.8143598Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8143782Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8143851Z res = mod(**inputs) 2025-12-04T09:39:41.8144088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8144164Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8144399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8144458Z outputs = block( 2025-12-04T09:39:41.8144671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8144747Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8144970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8145043Z return func(*args, **kwargs) 2025-12-04T09:39:41.8145270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8145352Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8145586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8145651Z return func(*args, **kwargs) 2025-12-04T09:39:41.8145887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.8146060Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.8146265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8146381Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8146384Z 2025-12-04T09:39:41.8146464Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.8146570Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8146753Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8146860Z res = mod(**inputs) 2025-12-04T09:39:41.8147106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8147185Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8147415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8147517Z outputs = block( 2025-12-04T09:39:41.8147725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8147806Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8148031Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8148097Z return func(*args, **kwargs) 2025-12-04T09:39:41.8148340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8148424Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8148649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8148723Z return func(*args, **kwargs) 2025-12-04T09:39:41.8148955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:41.8149058Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:41.8149334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:41.8149458Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:41.8149461Z 2025-12-04T09:39:41.8149571Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8149757Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8149832Z res = mod(**inputs) 2025-12-04T09:39:41.8150070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8150147Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8150386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8150449Z outputs = block( 2025-12-04T09:39:41.8150658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8150740Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8150972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8151046Z return func(*args, **kwargs) 2025-12-04T09:39:41.8151279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8151362Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8151596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8151660Z return func(*args, **kwargs) 2025-12-04T09:39:41.8151902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8151980Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8152185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8152301Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8152304Z 2025-12-04T09:39:41.8152403Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8152621Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8152692Z res = mod(**inputs) 2025-12-04T09:39:41.8152925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8153008Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8153269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8153330Z outputs = block( 2025-12-04T09:39:41.8153544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8153616Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8153842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8153916Z return func(*args, **kwargs) 2025-12-04T09:39:41.8154146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8154238Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8154461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8154530Z return func(*args, **kwargs) 2025-12-04T09:39:41.8154766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8154842Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8155049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8155158Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8155161Z 2025-12-04T09:39:41.8155259Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8155452Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8155513Z res = mod(**inputs) 2025-12-04T09:39:41.8155745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8155828Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8156061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8156126Z outputs = block( 2025-12-04T09:39:41.8156332Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8156406Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8156642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8156706Z return func(*args, **kwargs) 2025-12-04T09:39:41.8156944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8157041Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8157268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8157354Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8157555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8157662Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8157666Z 2025-12-04T09:39:41.8157768Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8157952Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8158019Z res = mod(**inputs) 2025-12-04T09:39:41.8158281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8158361Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8158600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8158701Z outputs = block( 2025-12-04T09:39:41.8158906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8158988Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8159212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8159286Z return func(*args, **kwargs) 2025-12-04T09:39:41.8159516Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8159616Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8159865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8159941Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8160154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8160264Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8160268Z 2025-12-04T09:39:41.8160366Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8160556Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8160617Z res = mod(**inputs) 2025-12-04T09:39:41.8160853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8160943Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8161174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8161241Z outputs = block( 2025-12-04T09:39:41.8161491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8161626Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8161877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8161944Z return func(*args, **kwargs) 2025-12-04T09:39:41.8162192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8162294Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8162540Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:41.8162628Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:41.8162841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:41.8163025Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:41.8163040Z 2025-12-04T09:39:41.8163146Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8163381Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8163463Z res = mod(**inputs) 2025-12-04T09:39:41.8163703Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8163786Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8164067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8164131Z outputs = block( 2025-12-04T09:39:41.8164349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8164425Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8164712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8164818Z return func(*args, **kwargs) 2025-12-04T09:39:41.8165051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8165146Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8165391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8165474Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8165691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8165802Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8165805Z 2025-12-04T09:39:41.8165904Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8166099Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8166163Z res = mod(**inputs) 2025-12-04T09:39:41.8166411Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8166489Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8166727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8166797Z outputs = block( 2025-12-04T09:39:41.8167011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8167087Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8167326Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8167392Z return func(*args, **kwargs) 2025-12-04T09:39:41.8167636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8167732Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8167972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8168060Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8168271Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8168382Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8168393Z 2025-12-04T09:39:41.8168492Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8168680Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8168748Z res = mod(**inputs) 2025-12-04T09:39:41.8168985Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8169066Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8169310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8169370Z outputs = block( 2025-12-04T09:39:41.8169587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8169661Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8169930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8170008Z return func(*args, **kwargs) 2025-12-04T09:39:41.8170243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 451, in forward 2025-12-04T09:39:41.8170345Z hidden_states = residual + feed_forward_hidden_states 2025-12-04T09:39:41.8170395Z 2025-12-04T09:39:41.8170496Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8170683Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8170752Z res = mod(**inputs) 2025-12-04T09:39:41.8170991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8171069Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8171315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8171375Z outputs = block( 2025-12-04T09:39:41.8171591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8171666Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8171897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8171973Z return func(*args, **kwargs) 2025-12-04T09:39:41.8172220Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8172304Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8172543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8172609Z return func(*args, **kwargs) 2025-12-04T09:39:41.8172856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.8173033Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.8173385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8173517Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8173521Z 2025-12-04T09:39:41.8173621Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8173820Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8173884Z res = mod(**inputs) 2025-12-04T09:39:41.8174124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8174213Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8174452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8174514Z outputs = block( 2025-12-04T09:39:41.8174736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8174816Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8175057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8175125Z return func(*args, **kwargs) 2025-12-04T09:39:41.8175363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8175459Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8175694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8175796Z return func(*args, **kwargs) 2025-12-04T09:39:41.8176047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.8176226Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.8176444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8176585Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8176588Z 2025-12-04T09:39:41.8176669Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.8176776Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8176969Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8177038Z res = mod(**inputs) 2025-12-04T09:39:41.8177283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8177361Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8177608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8177667Z outputs = block( 2025-12-04T09:39:41.8177872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8177954Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8178178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8178251Z return func(*args, **kwargs) 2025-12-04T09:39:41.8178478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8178559Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8178791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8178855Z return func(*args, **kwargs) 2025-12-04T09:39:41.8179092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:41.8179187Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:41.8179460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:41.8179588Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:41.8179591Z 2025-12-04T09:39:41.8179689Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8179873Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8179941Z res = mod(**inputs) 2025-12-04T09:39:41.8180178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8180261Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8180491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8180554Z outputs = block( 2025-12-04T09:39:41.8180767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8180841Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8181074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8181140Z return func(*args, **kwargs) 2025-12-04T09:39:41.8181369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8181493Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8181723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8181786Z return func(*args, **kwargs) 2025-12-04T09:39:41.8182025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8182149Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8182363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8182472Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8182475Z 2025-12-04T09:39:41.8182572Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8182766Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8182829Z res = mod(**inputs) 2025-12-04T09:39:41.8183074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8183162Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8183400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8183472Z outputs = block( 2025-12-04T09:39:41.8183683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8183759Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8184002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8184068Z return func(*args, **kwargs) 2025-12-04T09:39:41.8184309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8184395Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8184635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8184705Z return func(*args, **kwargs) 2025-12-04T09:39:41.8184936Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8185015Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8185226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8185334Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8185337Z 2025-12-04T09:39:41.8185443Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8185626Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8185687Z res = mod(**inputs) 2025-12-04T09:39:41.8185928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8186004Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8186238Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8186301Z outputs = block( 2025-12-04T09:39:41.8186507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8186608Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8186833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8186896Z return func(*args, **kwargs) 2025-12-04T09:39:41.8187171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8187271Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8187512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8187589Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8187795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8187956Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8187959Z 2025-12-04T09:39:41.8188057Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8188250Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8188310Z res = mod(**inputs) 2025-12-04T09:39:41.8188553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8188643Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8188880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8188943Z outputs = block( 2025-12-04T09:39:41.8189172Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8189248Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8189479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8189546Z return func(*args, **kwargs) 2025-12-04T09:39:41.8189779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8189886Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8190124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8190205Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8190426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8190534Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8190540Z 2025-12-04T09:39:41.8190645Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8190829Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8190889Z res = mod(**inputs) 2025-12-04T09:39:41.8191130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8191208Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8191445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8191505Z outputs = block( 2025-12-04T09:39:41.8191712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8191792Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8192017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8192084Z return func(*args, **kwargs) 2025-12-04T09:39:41.8192318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8192415Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8192650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:41.8192727Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:41.8192958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:41.8193139Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:41.8193142Z 2025-12-04T09:39:41.8193240Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8193460Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8193522Z res = mod(**inputs) 2025-12-04T09:39:41.8193755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8193842Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8194070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8194130Z outputs = block( 2025-12-04T09:39:41.8194346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8194422Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8194653Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8194719Z return func(*args, **kwargs) 2025-12-04T09:39:41.8194950Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8195054Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8195290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8195370Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8195581Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8195691Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8195694Z 2025-12-04T09:39:41.8195798Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8195980Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8196039Z res = mod(**inputs) 2025-12-04T09:39:41.8196283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8196359Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8196593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8196652Z outputs = block( 2025-12-04T09:39:41.8196858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8196940Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8197162Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8197227Z return func(*args, **kwargs) 2025-12-04T09:39:41.8197461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8197558Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8197791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8197871Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8198071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8198186Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8198189Z 2025-12-04T09:39:41.8198315Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8198514Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8198574Z res = mod(**inputs) 2025-12-04T09:39:41.8198805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8198920Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8199149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8199209Z outputs = block( 2025-12-04T09:39:41.8199425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8199499Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8199730Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8199797Z return func(*args, **kwargs) 2025-12-04T09:39:41.8200028Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8200119Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8200343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8200408Z return func(*args, **kwargs) 2025-12-04T09:39:41.8200642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.8200813Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.8201027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8201135Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8201141Z 2025-12-04T09:39:41.8201239Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8201462Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8201531Z res = mod(**inputs) 2025-12-04T09:39:41.8201886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8201979Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8202250Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8202327Z outputs = block( 2025-12-04T09:39:41.8202564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8202649Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8202903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8202969Z return func(*args, **kwargs) 2025-12-04T09:39:41.8203210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8203305Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8203573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8203656Z return func(*args, **kwargs) 2025-12-04T09:39:41.8203929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.8204133Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.8204367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8204543Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8204547Z 2025-12-04T09:39:41.8204645Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.8204758Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8204982Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8205092Z res = mod(**inputs) 2025-12-04T09:39:41.8205372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8205467Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8205743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8205811Z outputs = block( 2025-12-04T09:39:41.8206064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8206153Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8206425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8206499Z return func(*args, **kwargs) 2025-12-04T09:39:41.8206779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8206886Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8207156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8207231Z return func(*args, **kwargs) 2025-12-04T09:39:41.8207515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:41.8207623Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:41.8207951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:41.8208092Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:41.8208096Z 2025-12-04T09:39:41.8208207Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8208431Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8208503Z res = mod(**inputs) 2025-12-04T09:39:41.8208783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8208872Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8209147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8209232Z outputs = block( 2025-12-04T09:39:41.8209443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8209516Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8209752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8209816Z return func(*args, **kwargs) 2025-12-04T09:39:41.8210054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8210141Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8210367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8210439Z return func(*args, **kwargs) 2025-12-04T09:39:41.8210670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8210747Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8210986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8211098Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8211101Z 2025-12-04T09:39:41.8211205Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8211390Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8211479Z res = mod(**inputs) 2025-12-04T09:39:41.8211723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8211805Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8212044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8212104Z outputs = block( 2025-12-04T09:39:41.8212314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8212395Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8212620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8212685Z return func(*args, **kwargs) 2025-12-04T09:39:41.8212928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8213010Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8213244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8213308Z return func(*args, **kwargs) 2025-12-04T09:39:41.8213540Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8213625Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8213828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8213943Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8213946Z 2025-12-04T09:39:41.8214044Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8214230Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8214297Z res = mod(**inputs) 2025-12-04T09:39:41.8214533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8214609Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8214845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8214904Z outputs = block( 2025-12-04T09:39:41.8215120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8215194Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8215418Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8215492Z return func(*args, **kwargs) 2025-12-04T09:39:41.8215725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8215824Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8216061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8216140Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8216349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8216495Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8216498Z 2025-12-04T09:39:41.8216599Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8216790Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8216852Z res = mod(**inputs) 2025-12-04T09:39:41.8217127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8217204Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8217434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8217503Z outputs = block( 2025-12-04T09:39:41.8217710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8217787Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8218020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8218086Z return func(*args, **kwargs) 2025-12-04T09:39:41.8218323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8218424Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8218651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8218736Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8218939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8219056Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8219059Z 2025-12-04T09:39:41.8219160Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8219343Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8219411Z res = mod(**inputs) 2025-12-04T09:39:41.8219644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8219724Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8219960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8220020Z outputs = block( 2025-12-04T09:39:41.8220234Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8220308Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8220532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8220608Z return func(*args, **kwargs) 2025-12-04T09:39:41.8220839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8220934Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8221186Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:41.8221263Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:41.8221470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:41.8221638Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:41.8221641Z 2025-12-04T09:39:41.8221740Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8221930Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8222022Z res = mod(**inputs) 2025-12-04T09:39:41.8222263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8222340Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8222568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8222667Z outputs = block( 2025-12-04T09:39:41.8222883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8222958Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8223199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8223264Z return func(*args, **kwargs) 2025-12-04T09:39:41.8223511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8223607Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8223846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8223938Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8224149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8224268Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8224271Z 2025-12-04T09:39:41.8224368Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8224559Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8224628Z res = mod(**inputs) 2025-12-04T09:39:41.8224870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8224946Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8225189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8225250Z outputs = block( 2025-12-04T09:39:41.8225470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8225548Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8225779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8225851Z return func(*args, **kwargs) 2025-12-04T09:39:41.8226087Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8226189Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8226424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8226505Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8226720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8226830Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8226835Z 2025-12-04T09:39:41.8226933Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8227129Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8227192Z res = mod(**inputs) 2025-12-04T09:39:41.8227438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8227514Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8227790Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8227860Z outputs = block( 2025-12-04T09:39:41.8228068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8228151Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8228409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8228476Z return func(*args, **kwargs) 2025-12-04T09:39:41.8228707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 451, in forward 2025-12-04T09:39:41.8228808Z hidden_states = residual + feed_forward_hidden_states 2025-12-04T09:39:41.8228812Z 2025-12-04T09:39:41.8228907Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8229100Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8229161Z res = mod(**inputs) 2025-12-04T09:39:41.8229398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8229475Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8229702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8229774Z outputs = block( 2025-12-04T09:39:41.8229982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8230057Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8230441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8230513Z return func(*args, **kwargs) 2025-12-04T09:39:41.8230762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8230849Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8231076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8231149Z return func(*args, **kwargs) 2025-12-04T09:39:41.8231384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.8231567Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.8231773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8231885Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8231888Z 2025-12-04T09:39:41.8231996Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8232186Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8232249Z res = mod(**inputs) 2025-12-04T09:39:41.8232494Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8232573Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8232815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8232876Z outputs = block( 2025-12-04T09:39:41.8233085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8233171Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8233410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8233486Z return func(*args, **kwargs) 2025-12-04T09:39:41.8233773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8233861Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8234099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8234213Z return func(*args, **kwargs) 2025-12-04T09:39:41.8234447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:41.8234627Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:41.8234834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8234951Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8234954Z 2025-12-04T09:39:41.8235037Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.8235136Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8235333Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8235395Z res = mod(**inputs) 2025-12-04T09:39:41.8235647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8235727Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8235957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8236023Z outputs = block( 2025-12-04T09:39:41.8236228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8236304Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8236541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8236606Z return func(*args, **kwargs) 2025-12-04T09:39:41.8236842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8236923Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8237151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8237224Z return func(*args, **kwargs) 2025-12-04T09:39:41.8237453Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:41.8237545Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:41.8237825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:41.8237949Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:41.8237953Z 2025-12-04T09:39:41.8238058Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8238242Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8238303Z res = mod(**inputs) 2025-12-04T09:39:41.8238545Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8238622Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8238857Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8238918Z outputs = block( 2025-12-04T09:39:41.8239124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8239207Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8239465Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8239532Z return func(*args, **kwargs) 2025-12-04T09:39:41.8239773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8239888Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8240121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8240187Z return func(*args, **kwargs) 2025-12-04T09:39:41.8240417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8240502Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8240708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8240824Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8240827Z 2025-12-04T09:39:41.8240924Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8241108Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8241176Z res = mod(**inputs) 2025-12-04T09:39:41.8241420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8241498Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8241792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8241859Z outputs = block( 2025-12-04T09:39:41.8242081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8242161Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8242398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8242474Z return func(*args, **kwargs) 2025-12-04T09:39:41.8242716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:41.8242804Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:41.8243045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8243113Z return func(*args, **kwargs) 2025-12-04T09:39:41.8243365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:41.8243445Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:41.8243668Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8243787Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8243790Z 2025-12-04T09:39:41.8243890Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8244083Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8244153Z res = mod(**inputs) 2025-12-04T09:39:41.8244392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8244480Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8244715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8244777Z outputs = block( 2025-12-04T09:39:41.8244993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8245099Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8245381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8245447Z return func(*args, **kwargs) 2025-12-04T09:39:41.8245682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8245831Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8246072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8246159Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8246368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8246479Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8246482Z 2025-12-04T09:39:41.8246590Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8246781Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8246845Z res = mod(**inputs) 2025-12-04T09:39:41.8247093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8247174Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8247418Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8247479Z outputs = block( 2025-12-04T09:39:41.8247691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8247774Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8248010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8248078Z return func(*args, **kwargs) 2025-12-04T09:39:41.8248320Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8248419Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8248665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:41.8248747Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:41.8248954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8249073Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8249077Z 2025-12-04T09:39:41.8249176Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8249374Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8249436Z res = mod(**inputs) 2025-12-04T09:39:41.8249674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8249758Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8249995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8250058Z outputs = block( 2025-12-04T09:39:41.8250275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8250350Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8250588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8250655Z return func(*args, **kwargs) 2025-12-04T09:39:41.8250921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8251031Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8251270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:41.8251354Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:41.8251586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:41.8251760Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:41.8251763Z 2025-12-04T09:39:41.8251870Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8252058Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8252121Z res = mod(**inputs) 2025-12-04T09:39:41.8252374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8252455Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8252700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8252761Z outputs = block( 2025-12-04T09:39:41.8252975Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8253058Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8253288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8253363Z return func(*args, **kwargs) 2025-12-04T09:39:41.8253600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8253700Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8253942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8254025Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8254233Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8254355Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8254358Z 2025-12-04T09:39:41.8254457Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8254653Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8254715Z res = mod(**inputs) 2025-12-04T09:39:41.8254956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:41.8255042Z transformer_outputs = self.transformer( 2025-12-04T09:39:41.8255279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:41.8255346Z outputs = block( 2025-12-04T09:39:41.8255558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:41.8255631Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:41.8255872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:41.8255949Z return func(*args, **kwargs) 2025-12-04T09:39:41.8256178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:41.8256279Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:41.8256507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:41.8256622Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:41.8256827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:41.8256936Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:41.8256939Z 2025-12-04T09:39:41.8257026Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.8257133Z cudagraph partition due to non gpu ops 2025-12-04T09:39:41.8257232Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8257422Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8257483Z res = mod(**inputs) 2025-12-04T09:39:41.8257724Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1422, in forward 2025-12-04T09:39:41.8257862Z loss = loss_fct(pooled_logits.view(-1, self.num_labels), labels.view(-1)) 2025-12-04T09:39:41.8257866Z 2025-12-04T09:39:41.8257965Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:41.8258156Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:41.8258218Z res = mod(**inputs) 2025-12-04T09:39:41.8258460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1422, in forward 2025-12-04T09:39:41.8258596Z loss = loss_fct(pooled_logits.view(-1, self.num_labels), labels.view(-1)) 2025-12-04T09:39:41.8258600Z 2025-12-04T09:39:56.7898285Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.7903306Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.7905256Z res = mod(**inputs) 2025-12-04T09:39:56.7911325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.7913832Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.7914414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 873, in forward 2025-12-04T09:39:56.7918602Z causal_mask = create_causal_mask( 2025-12-04T09:39:56.7919100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 788, in create_causal_mask 2025-12-04T09:39:56.7919739Z early_exit, attention_mask, packed_sequence_mask, kv_length, kv_offset = _preprocess_mask_arguments( 2025-12-04T09:39:56.7920278Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 740, in _preprocess_mask_arguments 2025-12-04T09:39:56.7920748Z packed_sequence_mask = find_packed_sequence_indices(position_ids) 2025-12-04T09:39:56.7921224Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 655, in find_packed_sequence_indices 2025-12-04T09:39:56.7921900Z first_dummy_value = position_ids[:, :1] - 1 # We just need the diff on this first value to be 1 2025-12-04T09:39:56.7922137Z 2025-12-04T09:39:56.7922231Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.7922504Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.7922914Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.7923272Z res = mod(**inputs) 2025-12-04T09:39:56.7923645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1394, in forward 2025-12-04T09:39:56.7924089Z last_non_pad_token = (token_indices * non_pad_mask).argmax(-1) 2025-12-04T09:39:56.7924268Z 2025-12-04T09:39:56.7924396Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.7924741Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.7925090Z res = mod(**inputs) 2025-12-04T09:39:56.7925785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.7926189Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.7926572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.7926943Z outputs = block( 2025-12-04T09:39:56.7927362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.7927722Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.7928113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.7928487Z return func(*args, **kwargs) 2025-12-04T09:39:56.7928853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.7929245Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.7929648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.7930024Z return func(*args, **kwargs) 2025-12-04T09:39:56.7930612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.7931107Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.7931559Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.7931967Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.7932138Z 2025-12-04T09:39:56.7932244Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.7932599Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.7932922Z res = mod(**inputs) 2025-12-04T09:39:56.7933279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.7933668Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.7934101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.7934497Z outputs = block( 2025-12-04T09:39:56.7934836Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.7935203Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.7935582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.7935998Z return func(*args, **kwargs) 2025-12-04T09:39:56.7936375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.7936787Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.7937174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.7937547Z return func(*args, **kwargs) 2025-12-04T09:39:56.7937908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.7938395Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.7938849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.7939236Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.7939414Z 2025-12-04T09:39:56.7939520Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.7939937Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.7940258Z res = mod(**inputs) 2025-12-04T09:39:56.7940609Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.7941004Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.7941392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 873, in forward 2025-12-04T09:39:56.7942625Z causal_mask = create_causal_mask( 2025-12-04T09:39:56.7942995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 788, in create_causal_mask 2025-12-04T09:39:56.7943494Z early_exit, attention_mask, packed_sequence_mask, kv_length, kv_offset = _preprocess_mask_arguments( 2025-12-04T09:39:56.7944011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 740, in _preprocess_mask_arguments 2025-12-04T09:39:56.7944460Z packed_sequence_mask = find_packed_sequence_indices(position_ids) 2025-12-04T09:39:56.7944913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 656, in find_packed_sequence_indices 2025-12-04T09:39:56.7945389Z position_diff = torch.diff(position_ids, prepend=first_dummy_value, dim=-1) 2025-12-04T09:39:56.7945599Z 2025-12-04T09:39:56.7945694Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.7945931Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.7946293Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.7946625Z res = mod(**inputs) 2025-12-04T09:39:56.7946992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.7947390Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.7947791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 873, in forward 2025-12-04T09:39:56.7948182Z causal_mask = create_causal_mask( 2025-12-04T09:39:56.7948560Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 788, in create_causal_mask 2025-12-04T09:39:56.7949049Z early_exit, attention_mask, packed_sequence_mask, kv_length, kv_offset = _preprocess_mask_arguments( 2025-12-04T09:39:56.7949565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 740, in _preprocess_mask_arguments 2025-12-04T09:39:56.7950006Z packed_sequence_mask = find_packed_sequence_indices(position_ids) 2025-12-04T09:39:56.7950456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 657, in find_packed_sequence_indices 2025-12-04T09:39:56.7950889Z packed_sequence_mask = (position_diff != 1).cumsum(-1) 2025-12-04T09:39:56.7951072Z 2025-12-04T09:39:56.7951185Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.7951550Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.7951904Z res = mod(**inputs) 2025-12-04T09:39:56.7952264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.7952661Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.7953041Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 873, in forward 2025-12-04T09:39:56.7953423Z causal_mask = create_causal_mask( 2025-12-04T09:39:56.7953797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 788, in create_causal_mask 2025-12-04T09:39:56.7954290Z early_exit, attention_mask, packed_sequence_mask, kv_length, kv_offset = _preprocess_mask_arguments( 2025-12-04T09:39:56.7954854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 740, in _preprocess_mask_arguments 2025-12-04T09:39:56.7955311Z packed_sequence_mask = find_packed_sequence_indices(position_ids) 2025-12-04T09:39:56.7955770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/masking_utils.py", line 657, in find_packed_sequence_indices 2025-12-04T09:39:56.7956245Z packed_sequence_mask = (position_diff != 1).cumsum(-1) 2025-12-04T09:39:56.7956412Z 2025-12-04T09:39:56.7956523Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.7956897Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.7957243Z res = mod(**inputs) 2025-12-04T09:39:56.7957625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.7958072Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.7958479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.7958868Z outputs = block( 2025-12-04T09:39:56.7959202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.7959602Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.7960033Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.7960438Z return func(*args, **kwargs) 2025-12-04T09:39:56.7960846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.7961284Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.7961787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.7962189Z return func(*args, **kwargs) 2025-12-04T09:39:56.7962592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:56.7963034Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:56.7963519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:56.7964027Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:56.7964213Z 2025-12-04T09:39:56.7964296Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.7964535Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.7964884Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.7965191Z res = mod(**inputs) 2025-12-04T09:39:56.7965542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.7965927Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.7966297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.7966657Z outputs = block( 2025-12-04T09:39:56.7966982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.7967354Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.7967718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.7968078Z return func(*args, **kwargs) 2025-12-04T09:39:56.7968441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.7968817Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.7969227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.7969592Z return func(*args, **kwargs) 2025-12-04T09:39:56.7969952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:56.7970335Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:56.7970802Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:56.7971270Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:56.7971445Z 2025-12-04T09:39:56.7971557Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.7972270Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.7972590Z res = mod(**inputs) 2025-12-04T09:39:56.7972953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.7973343Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.7973718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.7974083Z outputs = block( 2025-12-04T09:39:56.7974402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.7974751Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.7975201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.7975568Z return func(*args, **kwargs) 2025-12-04T09:39:56.7975934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.7976314Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.7976702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.7977106Z return func(*args, **kwargs) 2025-12-04T09:39:56.7977460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.7977843Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.7978193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.7978583Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.7978751Z 2025-12-04T09:39:56.7978853Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.7979205Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.7979518Z res = mod(**inputs) 2025-12-04T09:39:56.7979863Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.7980244Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.7980620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.7980978Z outputs = block( 2025-12-04T09:39:56.7981291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.7981644Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.7982013Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.7982371Z return func(*args, **kwargs) 2025-12-04T09:39:56.7982721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.7983104Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.7983543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.7983960Z return func(*args, **kwargs) 2025-12-04T09:39:56.7984356Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.7984765Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.7985114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.7985529Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.7985705Z 2025-12-04T09:39:56.7985810Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.7986176Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.7986489Z res = mod(**inputs) 2025-12-04T09:39:56.7986838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.7987225Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.7987606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.7987961Z outputs = block( 2025-12-04T09:39:56.7988297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.7988654Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.7989022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.7989384Z return func(*args, **kwargs) 2025-12-04T09:39:56.7989750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.7990164Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.7990559Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.7990940Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.7991293Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.7991685Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.7991852Z 2025-12-04T09:39:56.7991956Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.7992307Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.7992623Z res = mod(**inputs) 2025-12-04T09:39:56.7992967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.7993353Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.7993780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.7994140Z outputs = block( 2025-12-04T09:39:56.7994451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.7994807Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.7995174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.7995537Z return func(*args, **kwargs) 2025-12-04T09:39:56.7995890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.7996294Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.7996721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.7997095Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.7997443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.7997827Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.7997992Z 2025-12-04T09:39:56.7998144Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.7998490Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.7998803Z res = mod(**inputs) 2025-12-04T09:39:56.7999148Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.7999532Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.7999905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8000272Z outputs = block( 2025-12-04T09:39:56.8000588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8000932Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8001308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8001787Z return func(*args, **kwargs) 2025-12-04T09:39:56.8002191Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8002653Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8003036Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:56.8003406Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:56.8003742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:56.8004181Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:56.8004411Z 2025-12-04T09:39:56.8004514Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8004863Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8005164Z res = mod(**inputs) 2025-12-04T09:39:56.8005512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8005890Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8006259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8006604Z outputs = block( 2025-12-04T09:39:56.8006921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8007277Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8007641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8008012Z return func(*args, **kwargs) 2025-12-04T09:39:56.8008365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8008767Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8009158Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8009545Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8009904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8010332Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8010501Z 2025-12-04T09:39:56.8010603Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8010962Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8011265Z res = mod(**inputs) 2025-12-04T09:39:56.8011595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8011994Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8012357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8012712Z outputs = block( 2025-12-04T09:39:56.8013017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8013367Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8013734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8014085Z return func(*args, **kwargs) 2025-12-04T09:39:56.8014448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8014829Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8015214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8015585Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8015932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8016317Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8016483Z 2025-12-04T09:39:56.8016590Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8016935Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8017256Z res = mod(**inputs) 2025-12-04T09:39:56.8017593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8017966Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8018345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8018707Z outputs = block( 2025-12-04T09:39:56.8019020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8019364Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8019731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8020088Z return func(*args, **kwargs) 2025-12-04T09:39:56.8020441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8020824Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8021198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8021558Z return func(*args, **kwargs) 2025-12-04T09:39:56.8021911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8022386Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8022831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8023218Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8023384Z 2025-12-04T09:39:56.8023518Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8023869Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8024180Z res = mod(**inputs) 2025-12-04T09:39:56.8024519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8024927Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8025318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8025688Z outputs = block( 2025-12-04T09:39:56.8025994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8026341Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8026707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8027058Z return func(*args, **kwargs) 2025-12-04T09:39:56.8027413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8027796Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8028170Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8028528Z return func(*args, **kwargs) 2025-12-04T09:39:56.8028887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8029361Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8029803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8030351Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8030539Z 2025-12-04T09:39:56.8030623Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.8030862Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8031207Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8031521Z res = mod(**inputs) 2025-12-04T09:39:56.8031878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8032266Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8032641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8033000Z outputs = block( 2025-12-04T09:39:56.8033321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8033675Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8034038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8034400Z return func(*args, **kwargs) 2025-12-04T09:39:56.8034760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8035149Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8035528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8035886Z return func(*args, **kwargs) 2025-12-04T09:39:56.8036248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:56.8036633Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:56.8037149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:56.8037612Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:56.8037788Z 2025-12-04T09:39:56.8037897Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8038245Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8038609Z res = mod(**inputs) 2025-12-04T09:39:56.8038961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8039338Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8039719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8040079Z outputs = block( 2025-12-04T09:39:56.8040390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8040739Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8041106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8041463Z return func(*args, **kwargs) 2025-12-04T09:39:56.8041870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8042267Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8042652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8043020Z return func(*args, **kwargs) 2025-12-04T09:39:56.8043380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8043776Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8044142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8044538Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8044709Z 2025-12-04T09:39:56.8044809Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8045153Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8045463Z res = mod(**inputs) 2025-12-04T09:39:56.8045797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8046171Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8046534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8046879Z outputs = block( 2025-12-04T09:39:56.8047176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8047522Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8047878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8048221Z return func(*args, **kwargs) 2025-12-04T09:39:56.8048565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8048937Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8049301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8049643Z return func(*args, **kwargs) 2025-12-04T09:39:56.8049988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8050355Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8050718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8051095Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8051264Z 2025-12-04T09:39:56.8051363Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8051705Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8052046Z res = mod(**inputs) 2025-12-04T09:39:56.8052385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8052760Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8053127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8053471Z outputs = block( 2025-12-04T09:39:56.8053781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8054123Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8054475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8054828Z return func(*args, **kwargs) 2025-12-04T09:39:56.8055183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8055575Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8055958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8056326Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8056667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8057033Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8057205Z 2025-12-04T09:39:56.8057302Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8057641Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8057944Z res = mod(**inputs) 2025-12-04T09:39:56.8058274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8058649Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8059018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8059372Z outputs = block( 2025-12-04T09:39:56.8059674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8060013Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8060376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8060720Z return func(*args, **kwargs) 2025-12-04T09:39:56.8061071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8061458Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8061843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8062204Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8062538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8062912Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8063071Z 2025-12-04T09:39:56.8063175Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8063534Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8063842Z res = mod(**inputs) 2025-12-04T09:39:56.8064184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8064554Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8064923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8065300Z outputs = block( 2025-12-04T09:39:56.8065604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8065937Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8066295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8066646Z return func(*args, **kwargs) 2025-12-04T09:39:56.8066990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8067376Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8067763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:56.8068131Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:56.8068460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:56.8068889Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:56.8069106Z 2025-12-04T09:39:56.8069213Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8069556Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8069853Z res = mod(**inputs) 2025-12-04T09:39:56.8070195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8070569Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8070928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8071281Z outputs = block( 2025-12-04T09:39:56.8071589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8071931Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8072283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8072631Z return func(*args, **kwargs) 2025-12-04T09:39:56.8072982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8073358Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8073746Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8074118Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8074457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8074824Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8074992Z 2025-12-04T09:39:56.8075091Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8075432Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8075734Z res = mod(**inputs) 2025-12-04T09:39:56.8076064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8076473Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8076844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8077187Z outputs = block( 2025-12-04T09:39:56.8077493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8077864Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8078221Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8078563Z return func(*args, **kwargs) 2025-12-04T09:39:56.8078913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8079303Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8079680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8080048Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8080389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8080765Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8080928Z 2025-12-04T09:39:56.8081029Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8081370Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8081758Z res = mod(**inputs) 2025-12-04T09:39:56.8082130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8082538Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8082932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8083307Z outputs = block( 2025-12-04T09:39:56.8083608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8083960Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8084339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8084729Z return func(*args, **kwargs) 2025-12-04T09:39:56.8085106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8085518Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8085922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8086306Z return func(*args, **kwargs) 2025-12-04T09:39:56.8086682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8087188Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8087665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8088068Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8088253Z 2025-12-04T09:39:56.8088360Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8088732Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8089064Z res = mod(**inputs) 2025-12-04T09:39:56.8089422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8089827Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8090256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8090636Z outputs = block( 2025-12-04T09:39:56.8090959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8091332Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8091758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8092101Z return func(*args, **kwargs) 2025-12-04T09:39:56.8092450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8092823Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8093192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8093539Z return func(*args, **kwargs) 2025-12-04T09:39:56.8093889Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8094356Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8094784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8095164Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8095331Z 2025-12-04T09:39:56.8095409Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.8095640Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8095974Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8096283Z res = mod(**inputs) 2025-12-04T09:39:56.8096628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8097002Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8097362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8097710Z outputs = block( 2025-12-04T09:39:56.8098020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8098357Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8098719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8099069Z return func(*args, **kwargs) 2025-12-04T09:39:56.8099416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8099783Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8100151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8100502Z return func(*args, **kwargs) 2025-12-04T09:39:56.8100842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:56.8101226Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:56.8101651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:56.8102109Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:56.8102286Z 2025-12-04T09:39:56.8102386Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8102732Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8103038Z res = mod(**inputs) 2025-12-04T09:39:56.8103421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8103790Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8104153Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8104501Z outputs = block( 2025-12-04T09:39:56.8104830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8105172Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8105534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8105882Z return func(*args, **kwargs) 2025-12-04T09:39:56.8106223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8106597Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8106966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8107309Z return func(*args, **kwargs) 2025-12-04T09:39:56.8107658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8108028Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8108367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8108736Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8108903Z 2025-12-04T09:39:56.8109003Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8109342Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8109644Z res = mod(**inputs) 2025-12-04T09:39:56.8109976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8110346Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8110712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8111055Z outputs = block( 2025-12-04T09:39:56.8111362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8111704Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8112058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8112402Z return func(*args, **kwargs) 2025-12-04T09:39:56.8112748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8113122Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8113480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8113831Z return func(*args, **kwargs) 2025-12-04T09:39:56.8114177Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8114548Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8114876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8115251Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8115411Z 2025-12-04T09:39:56.8115518Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8115859Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8116155Z res = mod(**inputs) 2025-12-04T09:39:56.8116520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8116896Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8117253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8117629Z outputs = block( 2025-12-04T09:39:56.8117937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8118279Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8118632Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8118984Z return func(*args, **kwargs) 2025-12-04T09:39:56.8119342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8119741Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8120139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8120517Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8120869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8121253Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8121425Z 2025-12-04T09:39:56.8121526Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8121948Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8122275Z res = mod(**inputs) 2025-12-04T09:39:56.8122624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8123017Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8123397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8123795Z outputs = block( 2025-12-04T09:39:56.8124121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8124493Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8124864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8125219Z return func(*args, **kwargs) 2025-12-04T09:39:56.8125576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8125979Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8126371Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8126821Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8127167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8127553Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8127718Z 2025-12-04T09:39:56.8127823Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8128174Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8128480Z res = mod(**inputs) 2025-12-04T09:39:56.8128821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8129239Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8129614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8130006Z outputs = block( 2025-12-04T09:39:56.8130460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8130820Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8131190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8131612Z return func(*args, **kwargs) 2025-12-04T09:39:56.8132012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8132415Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8132816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:56.8133197Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:56.8133539Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:56.8133987Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:56.8134213Z 2025-12-04T09:39:56.8134326Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8134672Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8134994Z res = mod(**inputs) 2025-12-04T09:39:56.8135345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8135732Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8136106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8136471Z outputs = block( 2025-12-04T09:39:56.8136796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8137143Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8137520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8137883Z return func(*args, **kwargs) 2025-12-04T09:39:56.8138244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8138639Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8139037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8139419Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8139768Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8140156Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8140324Z 2025-12-04T09:39:56.8140424Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8157038Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8157378Z res = mod(**inputs) 2025-12-04T09:39:56.8157766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8158181Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8158567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8158939Z outputs = block( 2025-12-04T09:39:56.8159263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8159622Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8160133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8160505Z return func(*args, **kwargs) 2025-12-04T09:39:56.8160874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8161273Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8161825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8162254Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8162633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8163054Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8163242Z 2025-12-04T09:39:56.8163352Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8163735Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8164075Z res = mod(**inputs) 2025-12-04T09:39:56.8164448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8164863Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8165284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8165668Z outputs = block( 2025-12-04T09:39:56.8166010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8166393Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8166798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8167183Z return func(*args, **kwargs) 2025-12-04T09:39:56.8167580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 451, in forward 2025-12-04T09:39:56.8168018Z hidden_states = residual + feed_forward_hidden_states 2025-12-04T09:39:56.8168185Z 2025-12-04T09:39:56.8168305Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8168677Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8169025Z res = mod(**inputs) 2025-12-04T09:39:56.8169398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8169812Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8170225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8170613Z outputs = block( 2025-12-04T09:39:56.8170959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8171336Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8171737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8172102Z return func(*args, **kwargs) 2025-12-04T09:39:56.8172455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8172842Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8173217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8173580Z return func(*args, **kwargs) 2025-12-04T09:39:56.8173930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8174446Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8174891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8175274Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8175439Z 2025-12-04T09:39:56.8175542Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8175918Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8176229Z res = mod(**inputs) 2025-12-04T09:39:56.8176565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8176946Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8177318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8177675Z outputs = block( 2025-12-04T09:39:56.8177980Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8178330Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8178693Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8179044Z return func(*args, **kwargs) 2025-12-04T09:39:56.8179397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8179778Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8180150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8180497Z return func(*args, **kwargs) 2025-12-04T09:39:56.8180854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8181321Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8181764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8182138Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8182313Z 2025-12-04T09:39:56.8182396Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.8182636Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8182971Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8183278Z res = mod(**inputs) 2025-12-04T09:39:56.8183620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8183997Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8184363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8184717Z outputs = block( 2025-12-04T09:39:56.8185028Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8185366Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8185733Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8186091Z return func(*args, **kwargs) 2025-12-04T09:39:56.8186444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8186817Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8187190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8187578Z return func(*args, **kwargs) 2025-12-04T09:39:56.8187933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:56.8188316Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:56.8188745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:56.8189258Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:56.8189431Z 2025-12-04T09:39:56.8189532Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8189873Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8190182Z res = mod(**inputs) 2025-12-04T09:39:56.8190516Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8190891Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8191257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8191608Z outputs = block( 2025-12-04T09:39:56.8191904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8192249Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8192607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8192948Z return func(*args, **kwargs) 2025-12-04T09:39:56.8193298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8193670Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8194036Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8194376Z return func(*args, **kwargs) 2025-12-04T09:39:56.8194721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8195090Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8195421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8195797Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8195966Z 2025-12-04T09:39:56.8196067Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8196403Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8196694Z res = mod(**inputs) 2025-12-04T09:39:56.8197033Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8197409Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8197774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8198117Z outputs = block( 2025-12-04T09:39:56.8198418Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8198760Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8199108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8199461Z return func(*args, **kwargs) 2025-12-04T09:39:56.8199806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8200176Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8200575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8200930Z return func(*args, **kwargs) 2025-12-04T09:39:56.8201278Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8201725Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8202083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8202531Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8202699Z 2025-12-04T09:39:56.8202813Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8203163Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8203485Z res = mod(**inputs) 2025-12-04T09:39:56.8203856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8204237Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8204599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8205003Z outputs = block( 2025-12-04T09:39:56.8205318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8205667Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8206040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8206400Z return func(*args, **kwargs) 2025-12-04T09:39:56.8206758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8207153Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8207552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8207928Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8208269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8208655Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8208831Z 2025-12-04T09:39:56.8208934Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8209287Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8209592Z res = mod(**inputs) 2025-12-04T09:39:56.8209941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8210325Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8210704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8211058Z outputs = block( 2025-12-04T09:39:56.8211372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8211727Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8212086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8212453Z return func(*args, **kwargs) 2025-12-04T09:39:56.8212817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8213218Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8213609Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8213985Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8214361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8214742Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8214913Z 2025-12-04T09:39:56.8215015Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8215370Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8215718Z res = mod(**inputs) 2025-12-04T09:39:56.8216071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8216464Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8216850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8217205Z outputs = block( 2025-12-04T09:39:56.8217511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8217863Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8218246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8218598Z return func(*args, **kwargs) 2025-12-04T09:39:56.8218956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8219351Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8219740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:56.8220106Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:56.8220449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:56.8220898Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:56.8221120Z 2025-12-04T09:39:56.8221232Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8221576Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8221886Z res = mod(**inputs) 2025-12-04T09:39:56.8222232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8222604Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8222978Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8223333Z outputs = block( 2025-12-04T09:39:56.8223643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8223985Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8224352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8224712Z return func(*args, **kwargs) 2025-12-04T09:39:56.8225063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8225460Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8225848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8226231Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8226575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8226960Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8227127Z 2025-12-04T09:39:56.8227267Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8227610Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8227905Z res = mod(**inputs) 2025-12-04T09:39:56.8228241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8228647Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8229006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8229354Z outputs = block( 2025-12-04T09:39:56.8229658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8229999Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8230487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8230848Z return func(*args, **kwargs) 2025-12-04T09:39:56.8231197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8231575Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8231962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8232343Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8232684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8233053Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8233225Z 2025-12-04T09:39:56.8233325Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8233664Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8233971Z res = mod(**inputs) 2025-12-04T09:39:56.8234300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8234671Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8235035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8235379Z outputs = block( 2025-12-04T09:39:56.8235683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8236028Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8236387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8236729Z return func(*args, **kwargs) 2025-12-04T09:39:56.8237082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8237453Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8237813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8238164Z return func(*args, **kwargs) 2025-12-04T09:39:56.8238512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8238984Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8239432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8239814Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8239978Z 2025-12-04T09:39:56.8240080Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8240514Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8240873Z res = mod(**inputs) 2025-12-04T09:39:56.8241231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8241673Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8242062Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8242498Z outputs = block( 2025-12-04T09:39:56.8242823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8243197Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8243584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8243946Z return func(*args, **kwargs) 2025-12-04T09:39:56.8244304Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8244691Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8245065Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8245418Z return func(*args, **kwargs) 2025-12-04T09:39:56.8245775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8246250Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8246696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8247071Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8247239Z 2025-12-04T09:39:56.8247320Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.8247559Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8247907Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8248211Z res = mod(**inputs) 2025-12-04T09:39:56.8248559Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8248944Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8249311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8249667Z outputs = block( 2025-12-04T09:39:56.8249978Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8250332Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8250695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8251058Z return func(*args, **kwargs) 2025-12-04T09:39:56.8251415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8251791Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8252169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8252531Z return func(*args, **kwargs) 2025-12-04T09:39:56.8252889Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:56.8253273Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:56.8253709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:56.8254216Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:56.8254395Z 2025-12-04T09:39:56.8254504Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8254857Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8255178Z res = mod(**inputs) 2025-12-04T09:39:56.8255513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8255913Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8256280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8256629Z outputs = block( 2025-12-04T09:39:56.8256929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8257262Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8257624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8257973Z return func(*args, **kwargs) 2025-12-04T09:39:56.8258314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8258682Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8259050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8259398Z return func(*args, **kwargs) 2025-12-04T09:39:56.8259739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8260107Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8260448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8260836Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8260996Z 2025-12-04T09:39:56.8261098Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8261442Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8261750Z res = mod(**inputs) 2025-12-04T09:39:56.8262085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8262461Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8262822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8263173Z outputs = block( 2025-12-04T09:39:56.8263469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8263813Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8264180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8264530Z return func(*args, **kwargs) 2025-12-04T09:39:56.8264883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8265265Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8265641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8265991Z return func(*args, **kwargs) 2025-12-04T09:39:56.8266356Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8266720Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8267057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8267454Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8267625Z 2025-12-04T09:39:56.8267725Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8268070Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8268364Z res = mod(**inputs) 2025-12-04T09:39:56.8268737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8269110Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8269488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8269831Z outputs = block( 2025-12-04T09:39:56.8270135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8270482Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8270859Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8271219Z return func(*args, **kwargs) 2025-12-04T09:39:56.8271584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8271978Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8272358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8272726Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8273061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8273438Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8273597Z 2025-12-04T09:39:56.8273697Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8274051Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8274352Z res = mod(**inputs) 2025-12-04T09:39:56.8274686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8275056Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8275422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8275765Z outputs = block( 2025-12-04T09:39:56.8276072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8276418Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8276776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8277124Z return func(*args, **kwargs) 2025-12-04T09:39:56.8277473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8277860Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8278235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8278604Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8278941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8279316Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8279477Z 2025-12-04T09:39:56.8279577Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8279918Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8280222Z res = mod(**inputs) 2025-12-04T09:39:56.8280605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8280981Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8281351Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8281810Z outputs = block( 2025-12-04T09:39:56.8282116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8282467Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8282837Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8283209Z return func(*args, **kwargs) 2025-12-04T09:39:56.8283567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8283980Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8284383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:56.8284766Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:56.8285115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:56.8285568Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:56.8285789Z 2025-12-04T09:39:56.8285898Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8286240Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8286551Z res = mod(**inputs) 2025-12-04T09:39:56.8286899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8287278Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8287644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8288002Z outputs = block( 2025-12-04T09:39:56.8288309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8288650Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8289012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8289370Z return func(*args, **kwargs) 2025-12-04T09:39:56.8289724Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8290111Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8290501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8290881Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8291224Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8291602Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8291776Z 2025-12-04T09:39:56.8291879Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8292226Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8292532Z res = mod(**inputs) 2025-12-04T09:39:56.8292881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8293264Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8293671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8294023Z outputs = block( 2025-12-04T09:39:56.8294336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8294687Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8295042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8295443Z return func(*args, **kwargs) 2025-12-04T09:39:56.8295800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8296197Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8296585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8296965Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8297359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8297731Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8297890Z 2025-12-04T09:39:56.8297989Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8298327Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8298630Z res = mod(**inputs) 2025-12-04T09:39:56.8298960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8299331Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8299692Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8300040Z outputs = block( 2025-12-04T09:39:56.8300339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8300681Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8301043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8301387Z return func(*args, **kwargs) 2025-12-04T09:39:56.8301736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 451, in forward 2025-12-04T09:39:56.8302125Z hidden_states = residual + feed_forward_hidden_states 2025-12-04T09:39:56.8302276Z 2025-12-04T09:39:56.8302382Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8302715Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8303016Z res = mod(**inputs) 2025-12-04T09:39:56.8303368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8303760Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8304119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8304472Z outputs = block( 2025-12-04T09:39:56.8304777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8305113Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8305469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8305822Z return func(*args, **kwargs) 2025-12-04T09:39:56.8306171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8306558Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8306969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8307333Z return func(*args, **kwargs) 2025-12-04T09:39:56.8307681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8307879Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8308116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8308235Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8308238Z 2025-12-04T09:39:56.8308337Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8308522Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8308591Z res = mod(**inputs) 2025-12-04T09:39:56.8308827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8308914Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8309142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8309205Z outputs = block( 2025-12-04T09:39:56.8309419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8309494Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8309720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8309794Z return func(*args, **kwargs) 2025-12-04T09:39:56.8310023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8310120Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8310342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8310409Z return func(*args, **kwargs) 2025-12-04T09:39:56.8310644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8310819Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8311031Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8311142Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8311145Z 2025-12-04T09:39:56.8311225Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.8311331Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8311520Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8311581Z res = mod(**inputs) 2025-12-04T09:39:56.8311820Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8311900Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8312134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8312194Z outputs = block( 2025-12-04T09:39:56.8312401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8312483Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8312707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8312780Z return func(*args, **kwargs) 2025-12-04T09:39:56.8313037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8313124Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8313354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8313451Z return func(*args, **kwargs) 2025-12-04T09:39:56.8313680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:56.8313780Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:56.8314051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:56.8314180Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:56.8314184Z 2025-12-04T09:39:56.8314284Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8314470Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8314538Z res = mod(**inputs) 2025-12-04T09:39:56.8314775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8314861Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8315092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8315152Z outputs = block( 2025-12-04T09:39:56.8315364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8315439Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8315664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8315739Z return func(*args, **kwargs) 2025-12-04T09:39:56.8315968Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8316056Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8316285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8316353Z return func(*args, **kwargs) 2025-12-04T09:39:56.8316588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8316666Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8316869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8316987Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8316990Z 2025-12-04T09:39:56.8317092Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8317284Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8317346Z res = mod(**inputs) 2025-12-04T09:39:56.8317576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8317667Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8317896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8317963Z outputs = block( 2025-12-04T09:39:56.8318168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8318243Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8318475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8318569Z return func(*args, **kwargs) 2025-12-04T09:39:56.8318800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8318891Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8319114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8319217Z return func(*args, **kwargs) 2025-12-04T09:39:56.8319445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8319522Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8319739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8319847Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8319851Z 2025-12-04T09:39:56.8319959Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8320151Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8320213Z res = mod(**inputs) 2025-12-04T09:39:56.8320452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8320533Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8320761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8320827Z outputs = block( 2025-12-04T09:39:56.8321034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8321116Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8321348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8321416Z return func(*args, **kwargs) 2025-12-04T09:39:56.8321726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8321836Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8322073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8322164Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8322377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8322503Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8322506Z 2025-12-04T09:39:56.8322612Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8322814Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8322892Z res = mod(**inputs) 2025-12-04T09:39:56.8323156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8323245Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8323488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8323554Z outputs = block( 2025-12-04T09:39:56.8323778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8323855Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8324092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8324170Z return func(*args, **kwargs) 2025-12-04T09:39:56.8324453Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8324560Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8324795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8324871Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8325120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8325230Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8325233Z 2025-12-04T09:39:56.8325338Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8325536Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8325599Z res = mod(**inputs) 2025-12-04T09:39:56.8325854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8325934Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8326179Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8326251Z outputs = block( 2025-12-04T09:39:56.8326469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8326554Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8326794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8326862Z return func(*args, **kwargs) 2025-12-04T09:39:56.8327111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8327214Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8327457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:56.8327545Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:56.8327753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:56.8327939Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:56.8327945Z 2025-12-04T09:39:56.8328047Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8328243Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8328315Z res = mod(**inputs) 2025-12-04T09:39:56.8328559Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8328646Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8328888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8328951Z outputs = block( 2025-12-04T09:39:56.8329168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8329245Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8329481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8329556Z return func(*args, **kwargs) 2025-12-04T09:39:56.8329797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8329904Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8330300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8330460Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8330683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8330799Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8330802Z 2025-12-04T09:39:56.8330913Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8331152Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8331218Z res = mod(**inputs) 2025-12-04T09:39:56.8331473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8331554Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8331795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8331866Z outputs = block( 2025-12-04T09:39:56.8332089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8332175Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8332413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8332486Z return func(*args, **kwargs) 2025-12-04T09:39:56.8332736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8332839Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8333089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8333175Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8333391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8333512Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8333515Z 2025-12-04T09:39:56.8333616Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8333810Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8333884Z res = mod(**inputs) 2025-12-04T09:39:56.8334129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8334220Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8334461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8334524Z outputs = block( 2025-12-04T09:39:56.8334749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8334828Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8335064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8335141Z return func(*args, **kwargs) 2025-12-04T09:39:56.8335381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8335477Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8335715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8335784Z return func(*args, **kwargs) 2025-12-04T09:39:56.8336039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8336222Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8336473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8336588Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8336592Z 2025-12-04T09:39:56.8336697Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8336899Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8337005Z res = mod(**inputs) 2025-12-04T09:39:56.8337246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8337323Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8337550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8337616Z outputs = block( 2025-12-04T09:39:56.8337823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8337896Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8338129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8338194Z return func(*args, **kwargs) 2025-12-04T09:39:56.8338428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8338513Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8338735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8338807Z return func(*args, **kwargs) 2025-12-04T09:39:56.8339044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8339219Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8339430Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8339538Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8339541Z 2025-12-04T09:39:56.8339627Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.8339727Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8339909Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8339978Z res = mod(**inputs) 2025-12-04T09:39:56.8340208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8340293Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8340521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8340582Z outputs = block( 2025-12-04T09:39:56.8340792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8340866Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8341090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8341163Z return func(*args, **kwargs) 2025-12-04T09:39:56.8341389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8341478Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8341701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8341765Z return func(*args, **kwargs) 2025-12-04T09:39:56.8342027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:56.8342124Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:56.8342404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:56.8342524Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:56.8342556Z 2025-12-04T09:39:56.8342655Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8342843Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8342903Z res = mod(**inputs) 2025-12-04T09:39:56.8343134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8343218Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8343450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8343514Z outputs = block( 2025-12-04T09:39:56.8343719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8343793Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8344022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8344088Z return func(*args, **kwargs) 2025-12-04T09:39:56.8344315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8344403Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8344623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8344694Z return func(*args, **kwargs) 2025-12-04T09:39:56.8344922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8344996Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8345205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8345313Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8345318Z 2025-12-04T09:39:56.8345422Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8345606Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8345667Z res = mod(**inputs) 2025-12-04T09:39:56.8345906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8345982Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8346209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8346276Z outputs = block( 2025-12-04T09:39:56.8346478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8346558Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8346783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8346849Z return func(*args, **kwargs) 2025-12-04T09:39:56.8347086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8347167Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8347395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8347459Z return func(*args, **kwargs) 2025-12-04T09:39:56.8347740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8347826Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8348028Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8348167Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8348170Z 2025-12-04T09:39:56.8348276Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8348460Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8348527Z res = mod(**inputs) 2025-12-04T09:39:56.8348759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8348837Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8349074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8349134Z outputs = block( 2025-12-04T09:39:56.8349340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8349421Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8349645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8349716Z return func(*args, **kwargs) 2025-12-04T09:39:56.8349942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8350040Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8350277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8350353Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8350562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8350672Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8350676Z 2025-12-04T09:39:56.8350774Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8350968Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8351028Z res = mod(**inputs) 2025-12-04T09:39:56.8351264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8351349Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8351580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8351645Z outputs = block( 2025-12-04T09:39:56.8351854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8351928Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8352159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8352225Z return func(*args, **kwargs) 2025-12-04T09:39:56.8352461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8352558Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8352789Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8352869Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8353098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8353208Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8353212Z 2025-12-04T09:39:56.8353317Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8353500Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8353567Z res = mod(**inputs) 2025-12-04T09:39:56.8353831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8353911Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8354148Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8354209Z outputs = block( 2025-12-04T09:39:56.8354415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8354499Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8354724Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8354795Z return func(*args, **kwargs) 2025-12-04T09:39:56.8355032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8355131Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8355366Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:56.8355441Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:56.8355646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:56.8355813Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:56.8355816Z 2025-12-04T09:39:56.8355918Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8356109Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8356169Z res = mod(**inputs) 2025-12-04T09:39:56.8356402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8356486Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8356714Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8356782Z outputs = block( 2025-12-04T09:39:56.8356987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8357063Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8357299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8357365Z return func(*args, **kwargs) 2025-12-04T09:39:56.8357603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8357698Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8357927Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8358016Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8358217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8358324Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8358333Z 2025-12-04T09:39:56.8358429Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8358639Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8358709Z res = mod(**inputs) 2025-12-04T09:39:56.8358942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8359019Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8359256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8359347Z outputs = block( 2025-12-04T09:39:56.8359563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8359637Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8359865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8359938Z return func(*args, **kwargs) 2025-12-04T09:39:56.8360174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8360272Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8360519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8360600Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8360822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8360933Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8360936Z 2025-12-04T09:39:56.8361035Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8361233Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8361294Z res = mod(**inputs) 2025-12-04T09:39:56.8361603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8361698Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8361968Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8362041Z outputs = block( 2025-12-04T09:39:56.8362266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8362351Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8362614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8362687Z return func(*args, **kwargs) 2025-12-04T09:39:56.8362946Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 451, in forward 2025-12-04T09:39:56.8363061Z hidden_states = residual + feed_forward_hidden_states 2025-12-04T09:39:56.8363068Z 2025-12-04T09:39:56.8363175Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8363389Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8363451Z res = mod(**inputs) 2025-12-04T09:39:56.8363700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8363791Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8364034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8364103Z outputs = block( 2025-12-04T09:39:56.8364323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8364401Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8364677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8364747Z return func(*args, **kwargs) 2025-12-04T09:39:56.8364992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8365078Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8365339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8365412Z return func(*args, **kwargs) 2025-12-04T09:39:56.8365656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8365831Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8366045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8366159Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8366163Z 2025-12-04T09:39:56.8366269Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8366459Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8366522Z res = mod(**inputs) 2025-12-04T09:39:56.8366772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8366852Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8367092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8367153Z outputs = block( 2025-12-04T09:39:56.8367364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8367447Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8367681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8367748Z return func(*args, **kwargs) 2025-12-04T09:39:56.8367993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8368080Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8368324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8368390Z return func(*args, **kwargs) 2025-12-04T09:39:56.8368627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8368808Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8369017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8369135Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8369138Z 2025-12-04T09:39:56.8369218Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.8369318Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8369513Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8369579Z res = mod(**inputs) 2025-12-04T09:39:56.8369816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8369900Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8370133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8370200Z outputs = block( 2025-12-04T09:39:56.8370449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8370527Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8370761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8370826Z return func(*args, **kwargs) 2025-12-04T09:39:56.8371086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8371177Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8371420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8371494Z return func(*args, **kwargs) 2025-12-04T09:39:56.8371735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:56.8371832Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:56.8372125Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:56.8372249Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:56.8372253Z 2025-12-04T09:39:56.8372355Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8372549Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8372611Z res = mod(**inputs) 2025-12-04T09:39:56.8372864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8372943Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8373183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8373250Z outputs = block( 2025-12-04T09:39:56.8373469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8373552Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8373788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8373856Z return func(*args, **kwargs) 2025-12-04T09:39:56.8374103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8374187Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8374427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8374494Z return func(*args, **kwargs) 2025-12-04T09:39:56.8374735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8374823Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8375034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8375146Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8375149Z 2025-12-04T09:39:56.8375255Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8375449Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8375529Z res = mod(**inputs) 2025-12-04T09:39:56.8375767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8375844Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8376086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8376171Z outputs = block( 2025-12-04T09:39:56.8376378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8376460Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8376689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8376791Z return func(*args, **kwargs) 2025-12-04T09:39:56.8377018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8377100Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8377334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8377400Z return func(*args, **kwargs) 2025-12-04T09:39:56.8377636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8377712Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8377912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8378025Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8378028Z 2025-12-04T09:39:56.8378128Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8378310Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8378377Z res = mod(**inputs) 2025-12-04T09:39:56.8378608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8378689Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8378916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8378977Z outputs = block( 2025-12-04T09:39:56.8379190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8379262Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8379490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8379559Z return func(*args, **kwargs) 2025-12-04T09:39:56.8379786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8379890Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8380117Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8380193Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8380404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8380511Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8380515Z 2025-12-04T09:39:56.8380618Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8380804Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8380868Z res = mod(**inputs) 2025-12-04T09:39:56.8381106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8381182Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8381415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8381474Z outputs = block( 2025-12-04T09:39:56.8381678Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8381784Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8382013Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8382078Z return func(*args, **kwargs) 2025-12-04T09:39:56.8382313Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8382954Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8383189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8383266Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8383468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8383583Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8383589Z 2025-12-04T09:39:56.8383685Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8383868Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8383936Z res = mod(**inputs) 2025-12-04T09:39:56.8384167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8384255Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8384481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8384540Z outputs = block( 2025-12-04T09:39:56.8384753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8384825Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8385054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8385120Z return func(*args, **kwargs) 2025-12-04T09:39:56.8385345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8385445Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8385672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:56.8385747Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:56.8385948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:56.8386114Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:56.8386117Z 2025-12-04T09:39:56.8386219Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8386404Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8386464Z res = mod(**inputs) 2025-12-04T09:39:56.8386700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8386776Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8387009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8387068Z outputs = block( 2025-12-04T09:39:56.8387271Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8387352Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8387573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8387637Z return func(*args, **kwargs) 2025-12-04T09:39:56.8387896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8387991Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8388228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8388344Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8388547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8388662Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8388665Z 2025-12-04T09:39:56.8388763Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8388951Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8389014Z res = mod(**inputs) 2025-12-04T09:39:56.8389245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8389329Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8389554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8389614Z outputs = block( 2025-12-04T09:39:56.8389829Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8389903Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8390135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8390200Z return func(*args, **kwargs) 2025-12-04T09:39:56.8390426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8390529Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8390759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8390838Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8391046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8391156Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8391159Z 2025-12-04T09:39:56.8391261Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8391443Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8391504Z res = mod(**inputs) 2025-12-04T09:39:56.8391743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8391822Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8392053Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8392112Z outputs = block( 2025-12-04T09:39:56.8392314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8392397Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8392623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8392690Z return func(*args, **kwargs) 2025-12-04T09:39:56.8392923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8393004Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8393235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8393326Z return func(*args, **kwargs) 2025-12-04T09:39:56.8393559Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8393738Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8393970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8394085Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8394088Z 2025-12-04T09:39:56.8394185Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8394369Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8394438Z res = mod(**inputs) 2025-12-04T09:39:56.8394674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8394753Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8394998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8395059Z outputs = block( 2025-12-04T09:39:56.8395271Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8395348Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8395571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8395645Z return func(*args, **kwargs) 2025-12-04T09:39:56.8395873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8395962Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8396187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8396255Z return func(*args, **kwargs) 2025-12-04T09:39:56.8396487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8396657Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8396861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8396977Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8396980Z 2025-12-04T09:39:56.8397058Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.8397162Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8397344Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8397406Z res = mod(**inputs) 2025-12-04T09:39:56.8397647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8397729Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8397962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8398032Z outputs = block( 2025-12-04T09:39:56.8398243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8398325Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8398554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8398620Z return func(*args, **kwargs) 2025-12-04T09:39:56.8398887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8398971Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8399207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8399275Z return func(*args, **kwargs) 2025-12-04T09:39:56.8399513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:56.8399644Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:56.8399926Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:56.8400053Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:56.8400062Z 2025-12-04T09:39:56.8400162Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8400353Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8400423Z res = mod(**inputs) 2025-12-04T09:39:56.8400664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8400745Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8400990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8401053Z outputs = block( 2025-12-04T09:39:56.8401275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8401352Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8401645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8401727Z return func(*args, **kwargs) 2025-12-04T09:39:56.8401967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8402052Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8402317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8402387Z return func(*args, **kwargs) 2025-12-04T09:39:56.8402650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8402734Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8402959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8403080Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8403084Z 2025-12-04T09:39:56.8403183Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8403380Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8403444Z res = mod(**inputs) 2025-12-04T09:39:56.8403685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8403773Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8404011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8404073Z outputs = block( 2025-12-04T09:39:56.8404292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8404370Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8404612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8404680Z return func(*args, **kwargs) 2025-12-04T09:39:56.8405016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8405112Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8405344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8405412Z return func(*args, **kwargs) 2025-12-04T09:39:56.8405684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8405761Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8405974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8406084Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8406088Z 2025-12-04T09:39:56.8406188Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8406385Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8406449Z res = mod(**inputs) 2025-12-04T09:39:56.8406696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8406778Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8407019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8407090Z outputs = block( 2025-12-04T09:39:56.8407301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8407378Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8407620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8407687Z return func(*args, **kwargs) 2025-12-04T09:39:56.8407931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8408029Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8408263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8408351Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8408558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8408667Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8408677Z 2025-12-04T09:39:56.8408775Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8408962Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8409031Z res = mod(**inputs) 2025-12-04T09:39:56.8409276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8409358Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8409598Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8409659Z outputs = block( 2025-12-04T09:39:56.8409883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8409958Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8410190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8410263Z return func(*args, **kwargs) 2025-12-04T09:39:56.8410496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8410623Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8410866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8410944Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8411155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8411301Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8411304Z 2025-12-04T09:39:56.8411403Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8411604Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8411666Z res = mod(**inputs) 2025-12-04T09:39:56.8411912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8411992Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8412229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8412297Z outputs = block( 2025-12-04T09:39:56.8412509Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8412585Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8412825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8412891Z return func(*args, **kwargs) 2025-12-04T09:39:56.8413132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8413228Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8413460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:56.8413547Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:56.8413749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:56.8413927Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:56.8413933Z 2025-12-04T09:39:56.8414031Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8414219Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8414289Z res = mod(**inputs) 2025-12-04T09:39:56.8414529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8414609Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8414854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8414915Z outputs = block( 2025-12-04T09:39:56.8415135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8415209Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8415440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8415519Z return func(*args, **kwargs) 2025-12-04T09:39:56.8415753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8415857Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8416093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8416175Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8416419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8416542Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8416545Z 2025-12-04T09:39:56.8416643Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8416835Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8416927Z res = mod(**inputs) 2025-12-04T09:39:56.8417174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8417251Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8417485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8417552Z outputs = block( 2025-12-04T09:39:56.8417775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8417849Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8418086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8418151Z return func(*args, **kwargs) 2025-12-04T09:39:56.8418395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8418495Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8418729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8418818Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8419030Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8419152Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8419155Z 2025-12-04T09:39:56.8419254Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8419448Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8419517Z res = mod(**inputs) 2025-12-04T09:39:56.8419761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8419845Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8420093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8420155Z outputs = block( 2025-12-04T09:39:56.8420377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8420453Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8420712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8420784Z return func(*args, **kwargs) 2025-12-04T09:39:56.8421021Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 451, in forward 2025-12-04T09:39:56.8421130Z hidden_states = residual + feed_forward_hidden_states 2025-12-04T09:39:56.8421136Z 2025-12-04T09:39:56.8421232Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8421419Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8421485Z res = mod(**inputs) 2025-12-04T09:39:56.8421723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8421799Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8422068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8422129Z outputs = block( 2025-12-04T09:39:56.8422342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8422414Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8422640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8422740Z return func(*args, **kwargs) 2025-12-04T09:39:56.8422972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8423056Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8423289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8423355Z return func(*args, **kwargs) 2025-12-04T09:39:56.8423598Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8423771Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8423978Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8424098Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8424101Z 2025-12-04T09:39:56.8424199Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8424387Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8424449Z res = mod(**inputs) 2025-12-04T09:39:56.8424681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8424767Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8424997Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8425057Z outputs = block( 2025-12-04T09:39:56.8425269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8425343Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8425575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8425640Z return func(*args, **kwargs) 2025-12-04T09:39:56.8425867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8425958Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8426184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8426256Z return func(*args, **kwargs) 2025-12-04T09:39:56.8426490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8426662Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8426870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8426982Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8426985Z 2025-12-04T09:39:56.8427067Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.8427163Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8427344Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8427411Z res = mod(**inputs) 2025-12-04T09:39:56.8427673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8427753Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8427986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8428046Z outputs = block( 2025-12-04T09:39:56.8428258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8428359Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8428584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8428656Z return func(*args, **kwargs) 2025-12-04T09:39:56.8428888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8428970Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8429202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8429268Z return func(*args, **kwargs) 2025-12-04T09:39:56.8429504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:56.8429597Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:56.8429875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:56.8430002Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:56.8430005Z 2025-12-04T09:39:56.8430102Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8430414Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8430478Z res = mod(**inputs) 2025-12-04T09:39:56.8430716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8430805Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8431037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8431101Z outputs = block( 2025-12-04T09:39:56.8431318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8431395Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8431637Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8431705Z return func(*args, **kwargs) 2025-12-04T09:39:56.8431936Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8432030Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8432258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8432325Z return func(*args, **kwargs) 2025-12-04T09:39:56.8432563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8432644Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8432856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8432968Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8432971Z 2025-12-04T09:39:56.8433074Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8433265Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8433329Z res = mod(**inputs) 2025-12-04T09:39:56.8433626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8433708Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8433937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8434045Z outputs = block( 2025-12-04T09:39:56.8434254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8434328Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8434562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8434626Z return func(*args, **kwargs) 2025-12-04T09:39:56.8434864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8434947Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8435168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8435242Z return func(*args, **kwargs) 2025-12-04T09:39:56.8435470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8435550Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8435761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8435869Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8435873Z 2025-12-04T09:39:56.8435981Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8436165Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8436227Z res = mod(**inputs) 2025-12-04T09:39:56.8436469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8436547Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8436782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8436844Z outputs = block( 2025-12-04T09:39:56.8437049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8437129Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8437357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8437421Z return func(*args, **kwargs) 2025-12-04T09:39:56.8437658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8437756Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8437989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8438065Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8438266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8438384Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8438387Z 2025-12-04T09:39:56.8438483Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8438670Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8438731Z res = mod(**inputs) 2025-12-04T09:39:56.8438964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8439097Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8439331Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8439392Z outputs = block( 2025-12-04T09:39:56.8439606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8439711Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8439947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8440016Z return func(*args, **kwargs) 2025-12-04T09:39:56.8440251Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8440357Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8440597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8440682Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8440889Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8441001Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8441007Z 2025-12-04T09:39:56.8441110Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8441305Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8441372Z res = mod(**inputs) 2025-12-04T09:39:56.8441687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8441777Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8442039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8442102Z outputs = block( 2025-12-04T09:39:56.8442325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8442412Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8442668Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8442739Z return func(*args, **kwargs) 2025-12-04T09:39:56.8443056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8443152Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8443392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:56.8443469Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:56.8443675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:56.8443850Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:56.8443854Z 2025-12-04T09:39:56.8443952Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8444153Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8444217Z res = mod(**inputs) 2025-12-04T09:39:56.8444475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8444568Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8444823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8444887Z outputs = block( 2025-12-04T09:39:56.8445152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8445234Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8445486Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8445557Z return func(*args, **kwargs) 2025-12-04T09:39:56.8445868Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8445980Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8446232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8446328Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8446549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8446669Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8446673Z 2025-12-04T09:39:56.8446786Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8446985Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8447050Z res = mod(**inputs) 2025-12-04T09:39:56.8447317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8447401Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8447655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8447718Z outputs = block( 2025-12-04T09:39:56.8447941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8448028Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8448273Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8448353Z return func(*args, **kwargs) 2025-12-04T09:39:56.8448604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8448711Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8448968Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8449054Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8449282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8449410Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8449413Z 2025-12-04T09:39:56.8449517Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8449729Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8449794Z res = mod(**inputs) 2025-12-04T09:39:56.8450052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8450143Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8450397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8450467Z outputs = block( 2025-12-04T09:39:56.8450692Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8450769Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8451019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8451118Z return func(*args, **kwargs) 2025-12-04T09:39:56.8451370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8451470Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8451712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8451815Z return func(*args, **kwargs) 2025-12-04T09:39:56.8452061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8452249Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8452457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8452563Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8452566Z 2025-12-04T09:39:56.8452666Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8452848Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8452906Z res = mod(**inputs) 2025-12-04T09:39:56.8453143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8453224Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8453450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8453515Z outputs = block( 2025-12-04T09:39:56.8453716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8453793Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8454018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8454081Z return func(*args, **kwargs) 2025-12-04T09:39:56.8454312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8454393Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8454612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8454680Z return func(*args, **kwargs) 2025-12-04T09:39:56.8454910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8455081Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8455285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8455393Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8455396Z 2025-12-04T09:39:56.8455474Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.8455568Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8455754Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8455817Z res = mod(**inputs) 2025-12-04T09:39:56.8456052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8456137Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8456369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8456428Z outputs = block( 2025-12-04T09:39:56.8456643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8456752Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8456983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8457049Z return func(*args, **kwargs) 2025-12-04T09:39:56.8457278Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8457397Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8457621Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8457686Z return func(*args, **kwargs) 2025-12-04T09:39:56.8457921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:56.8458014Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:56.8458294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:56.8458417Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:56.8458420Z 2025-12-04T09:39:56.8458519Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8458707Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8458771Z res = mod(**inputs) 2025-12-04T09:39:56.8459008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8459087Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8459317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8459385Z outputs = block( 2025-12-04T09:39:56.8459592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8459666Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8459894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8459960Z return func(*args, **kwargs) 2025-12-04T09:39:56.8460198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8460287Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8460517Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8460583Z return func(*args, **kwargs) 2025-12-04T09:39:56.8460811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8460887Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8461093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8461202Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8461206Z 2025-12-04T09:39:56.8461308Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8461490Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8461555Z res = mod(**inputs) 2025-12-04T09:39:56.8461795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8461876Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8462113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8462173Z outputs = block( 2025-12-04T09:39:56.8462409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8462492Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8462721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8462794Z return func(*args, **kwargs) 2025-12-04T09:39:56.8463039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8463154Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8463395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8463463Z return func(*args, **kwargs) 2025-12-04T09:39:56.8463705Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8463791Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8464009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8464130Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8464134Z 2025-12-04T09:39:56.8464245Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8464428Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8464499Z res = mod(**inputs) 2025-12-04T09:39:56.8464742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8464824Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8465074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8465135Z outputs = block( 2025-12-04T09:39:56.8465359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8465436Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8465673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8465748Z return func(*args, **kwargs) 2025-12-04T09:39:56.8465998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8466102Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8466335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8466411Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8466625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8466737Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8466740Z 2025-12-04T09:39:56.8466838Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8467028Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8467085Z res = mod(**inputs) 2025-12-04T09:39:56.8467321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8467399Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8467630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8467696Z outputs = block( 2025-12-04T09:39:56.8467905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8471729Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8472028Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8472100Z return func(*args, **kwargs) 2025-12-04T09:39:56.8472349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8472449Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8472701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8472786Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8472990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8473101Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8473113Z 2025-12-04T09:39:56.8473237Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8473425Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8473494Z res = mod(**inputs) 2025-12-04T09:39:56.8473726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8473805Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8474035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8474093Z outputs = block( 2025-12-04T09:39:56.8474299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8474374Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8474597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8474671Z return func(*args, **kwargs) 2025-12-04T09:39:56.8474903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8475000Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8475241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:56.8475319Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:56.8475518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:56.8475696Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:56.8475699Z 2025-12-04T09:39:56.8475797Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8475990Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8476055Z res = mod(**inputs) 2025-12-04T09:39:56.8476286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8476371Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8476599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8476666Z outputs = block( 2025-12-04T09:39:56.8476876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8476950Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8477178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8477245Z return func(*args, **kwargs) 2025-12-04T09:39:56.8477506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8477664Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8477907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8478000Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8478212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8478357Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8478360Z 2025-12-04T09:39:56.8478468Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8478661Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8478728Z res = mod(**inputs) 2025-12-04T09:39:56.8478971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8479057Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8479305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8479369Z outputs = block( 2025-12-04T09:39:56.8479587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8479674Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8479911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8479987Z return func(*args, **kwargs) 2025-12-04T09:39:56.8480229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8480329Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8480589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8480678Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8480908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8481030Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8481035Z 2025-12-04T09:39:56.8481138Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8481345Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8481410Z res = mod(**inputs) 2025-12-04T09:39:56.8481810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8481909Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8482162Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8482231Z outputs = block( 2025-12-04T09:39:56.8482451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8482527Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8482773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8482843Z return func(*args, **kwargs) 2025-12-04T09:39:56.8483101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 451, in forward 2025-12-04T09:39:56.8483208Z hidden_states = residual + feed_forward_hidden_states 2025-12-04T09:39:56.8483212Z 2025-12-04T09:39:56.8483311Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8483503Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8483633Z res = mod(**inputs) 2025-12-04T09:39:56.8483879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8483969Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8484208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8484294Z outputs = block( 2025-12-04T09:39:56.8484510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8484588Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8484832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8484901Z return func(*args, **kwargs) 2025-12-04T09:39:56.8485148Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8485243Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8485480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8485555Z return func(*args, **kwargs) 2025-12-04T09:39:56.8485799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8485986Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8486215Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8486332Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8486336Z 2025-12-04T09:39:56.8486444Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8486643Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8486708Z res = mod(**inputs) 2025-12-04T09:39:56.8486963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8487045Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8487288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8487356Z outputs = block( 2025-12-04T09:39:56.8487573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8487656Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8487891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8487961Z return func(*args, **kwargs) 2025-12-04T09:39:56.8488214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8488302Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8488544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8488613Z return func(*args, **kwargs) 2025-12-04T09:39:56.8488853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 313, in forward 2025-12-04T09:39:56.8489039Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-12-04T09:39:56.8489254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8489368Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8489399Z 2025-12-04T09:39:56.8489511Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.8489615Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8489812Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8489877Z res = mod(**inputs) 2025-12-04T09:39:56.8490119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8490227Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8490471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8490541Z outputs = block( 2025-12-04T09:39:56.8490759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8490837Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8491086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8491153Z return func(*args, **kwargs) 2025-12-04T09:39:56.8491399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8491491Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8491730Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8491805Z return func(*args, **kwargs) 2025-12-04T09:39:56.8492052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 345, in forward 2025-12-04T09:39:56.8492149Z attn_output, attn_weights = attention_interface( 2025-12-04T09:39:56.8492445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:39:56.8492579Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:39:56.8492583Z 2025-12-04T09:39:56.8492692Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8492887Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8492951Z res = mod(**inputs) 2025-12-04T09:39:56.8493219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8493303Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8493548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8493619Z outputs = block( 2025-12-04T09:39:56.8493838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8493926Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8494169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8494239Z return func(*args, **kwargs) 2025-12-04T09:39:56.8494488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8494575Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8494815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8494889Z return func(*args, **kwargs) 2025-12-04T09:39:56.8495134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8495220Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8495432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8495597Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8495600Z 2025-12-04T09:39:56.8495712Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8495906Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8495976Z res = mod(**inputs) 2025-12-04T09:39:56.8496237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8496320Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8496623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8496696Z outputs = block( 2025-12-04T09:39:56.8496906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8496993Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8497225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8497295Z return func(*args, **kwargs) 2025-12-04T09:39:56.8497530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 413, in forward 2025-12-04T09:39:56.8497613Z attn_output, self_attn_weights = self.attn( 2025-12-04T09:39:56.8497846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8497913Z return func(*args, **kwargs) 2025-12-04T09:39:56.8498147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 358, in forward 2025-12-04T09:39:56.8498231Z attn_output = self.c_proj(attn_output) 2025-12-04T09:39:56.8498437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8498559Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8498562Z 2025-12-04T09:39:56.8498662Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8498849Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8498917Z res = mod(**inputs) 2025-12-04T09:39:56.8499157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8499244Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8499479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8499540Z outputs = block( 2025-12-04T09:39:56.8499760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8499839Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8500069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8500145Z return func(*args, **kwargs) 2025-12-04T09:39:56.8500382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8500490Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8500731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8500809Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8501028Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8501139Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8501159Z 2025-12-04T09:39:56.8501304Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8501495Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8501557Z res = mod(**inputs) 2025-12-04T09:39:56.8501807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8501900Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8502134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8502202Z outputs = block( 2025-12-04T09:39:56.8502410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8502492Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8502721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8502790Z return func(*args, **kwargs) 2025-12-04T09:39:56.8503030Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8503128Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8503363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 374, in forward 2025-12-04T09:39:56.8503449Z hidden_states = self.c_fc(hidden_states) 2025-12-04T09:39:56.8503652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8503768Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8503771Z 2025-12-04T09:39:56.8503870Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8504058Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8504130Z res = mod(**inputs) 2025-12-04T09:39:56.8504366Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8504452Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8504685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8504748Z outputs = block( 2025-12-04T09:39:56.8504962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8505036Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8505265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8505339Z return func(*args, **kwargs) 2025-12-04T09:39:56.8505573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8505680Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8505916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 375, in forward 2025-12-04T09:39:56.8505994Z hidden_states = self.act(hidden_states) 2025-12-04T09:39:56.8506203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:39:56.8506380Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:39:56.8506383Z 2025-12-04T09:39:56.8506488Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8506675Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8506738Z res = mod(**inputs) 2025-12-04T09:39:56.8507026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8507107Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8507340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8507409Z outputs = block( 2025-12-04T09:39:56.8507623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8507722Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8507952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8508019Z return func(*args, **kwargs) 2025-12-04T09:39:56.8508262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8508362Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8508608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8508691Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8508900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8509020Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8509023Z 2025-12-04T09:39:56.8509122Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8509318Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8509386Z res = mod(**inputs) 2025-12-04T09:39:56.8509628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1365, in forward 2025-12-04T09:39:56.8509714Z transformer_outputs = self.transformer( 2025-12-04T09:39:56.8509949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 925, in forward 2025-12-04T09:39:56.8510009Z outputs = block( 2025-12-04T09:39:56.8510227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:39:56.8510303Z return super().__call__(*args, **kwargs) 2025-12-04T09:39:56.8510541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:39:56.8510608Z return func(*args, **kwargs) 2025-12-04T09:39:56.8510842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 449, in forward 2025-12-04T09:39:56.8510955Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-12-04T09:39:56.8511192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 376, in forward 2025-12-04T09:39:56.8511278Z hidden_states = self.c_proj(hidden_states) 2025-12-04T09:39:56.8511495Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 122, in forward 2025-12-04T09:39:56.8511606Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-12-04T09:39:56.8511610Z 2025-12-04T09:39:56.8511698Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.8511788Z cudagraph partition due to non gpu ops 2025-12-04T09:39:56.8511885Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8512073Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8512135Z res = mod(**inputs) 2025-12-04T09:39:56.8512369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1422, in forward 2025-12-04T09:39:56.8512515Z loss = loss_fct(pooled_logits.view(-1, self.num_labels), labels.view(-1)) 2025-12-04T09:39:56.8512538Z 2025-12-04T09:39:56.8512663Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:39:56.8512862Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:39:56.8512924Z res = mod(**inputs) 2025-12-04T09:39:56.8513157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1422, in forward 2025-12-04T09:39:56.8513322Z loss = loss_fct(pooled_logits.view(-1, self.num_labels), labels.view(-1)) 2025-12-04T09:39:56.8513326Z 2025-12-04T09:39:58.3113594Z Compilation time (from dynamo_timed): 24.018346449 2025-12-04T09:39:58.3113909Z pass 2025-12-04T09:39:58.3114197Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:39:58.3115015Z TIMING: _recursive_pre_grad_passes:0.06901 _recursive_joint_graph_passes:0.76371 _recursive_post_grad_passes:0.11621 async_compile.wait:0.69805 code_gen:10.633 inductor_compile:13.20946 backend_compile:19.82857 gc:0.00101 entire_frame_compile:24.01835 total_wall_time:24.01835 2025-12-04T09:39:58.3115890Z STATS: call_* op count: 1130 | FakeTensorMode.__torch_dispatch__:33845 | FakeTensor.__torch_dispatch__:7141 | ProxyTorchDispatchMode.__torch_dispatch__:4439 2025-12-04T09:39:58.3116366Z Dynamo produced 2 graphs covering 1130 ops with 0 graph breaks (0 unique) 2025-12-04T09:40:00.7548652Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:40:00.7550223Z import pynvml # type: ignore[import] 2025-12-04T09:40:03.9111906Z 2025-12-04T09:40:04.5735170Z loading model: 0it [00:00, ?it/s]WARNING:common:Model GoogleFnet supports float32 only 2025-12-04T09:40:04.6993840Z 2025-12-04T09:40:04.6997113Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:40:04.6997848Z WARNING:common:Model GoogleFnet supports float32 only 2025-12-04T09:40:04.7000579Z cpu eval GoogleFnet 2025-12-04T09:40:05.0985252Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:40:05.0987882Z WARNING:common:Model GoogleFnet supports float32 only 2025-12-04T09:40:05.2505040Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:40:05.4045646Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:40:11.0312613Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0313292Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0313687Z res = mod(**inputs) 2025-12-04T09:40:11.0314075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0314523Z outputs = self.fnet( 2025-12-04T09:40:11.0314891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0315279Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0315650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0316058Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0316431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0316795Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0317178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0317607Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0318478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0318874Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0319310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0319780Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0320013Z 2025-12-04T09:40:11.0320138Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0320538Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0320897Z res = mod(**inputs) 2025-12-04T09:40:11.0321291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0321870Z outputs = self.fnet( 2025-12-04T09:40:11.0322264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0322703Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0323110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0323505Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0323881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0324291Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0324683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0325082Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0325487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0325875Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0326258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0326673Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0326839Z 2025-12-04T09:40:11.0326948Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0327310Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0327628Z res = mod(**inputs) 2025-12-04T09:40:11.0328015Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0328482Z outputs = self.fnet( 2025-12-04T09:40:11.0328839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0329273Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0329657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0330055Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0330574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0330979Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0331364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0331766Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0332162Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0332554Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0332943Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0333468Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0333626Z 2025-12-04T09:40:11.0333734Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0334096Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0334423Z res = mod(**inputs) 2025-12-04T09:40:11.0334758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0335152Z outputs = self.fnet( 2025-12-04T09:40:11.0335491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0335860Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0336217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0336600Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0336961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0337308Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0337672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0338065Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0338456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0338826Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0339201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0339597Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0339746Z 2025-12-04T09:40:11.0339856Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0340201Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0340516Z res = mod(**inputs) 2025-12-04T09:40:11.0340860Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0341220Z outputs = self.fnet( 2025-12-04T09:40:11.0341566Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0341946Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0342313Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0342688Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0343045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0343400Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0343773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0344158Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0344542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0344921Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0345289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0345690Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0345845Z 2025-12-04T09:40:11.0345955Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0346295Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0346615Z res = mod(**inputs) 2025-12-04T09:40:11.0346983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0347339Z outputs = self.fnet( 2025-12-04T09:40:11.0347666Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0348045Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0348399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0348773Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0349114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0349462Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0349823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0350210Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0350577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0350942Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0351305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0351684Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0351835Z 2025-12-04T09:40:11.0351934Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0352284Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0352599Z res = mod(**inputs) 2025-12-04T09:40:11.0352932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0353299Z outputs = self.fnet( 2025-12-04T09:40:11.0353640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0354006Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0354360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0354745Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0355098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0355449Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0355823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0356215Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0356610Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0356982Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0357355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0357754Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0357903Z 2025-12-04T09:40:11.0358013Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0358353Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0358667Z res = mod(**inputs) 2025-12-04T09:40:11.0359007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0359366Z outputs = self.fnet( 2025-12-04T09:40:11.0359743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0360130Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0360489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0360863Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0361222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0361681Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0362075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0362513Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0362945Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0363323Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0363694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0364089Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0364237Z 2025-12-04T09:40:11.0364347Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0364698Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0365005Z res = mod(**inputs) 2025-12-04T09:40:11.0365365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0365737Z outputs = self.fnet( 2025-12-04T09:40:11.0366073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0366440Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0366804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0367183Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0367529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0367877Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0368250Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0368633Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0369020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0369397Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0369769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0370161Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0370321Z 2025-12-04T09:40:11.0370424Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0370777Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0371091Z res = mod(**inputs) 2025-12-04T09:40:11.0371426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0371797Z outputs = self.fnet( 2025-12-04T09:40:11.0372146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0372510Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0372875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0373289Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0373691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0374039Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0374425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0374839Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0375225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0375591Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0375964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0376357Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0376509Z 2025-12-04T09:40:11.0376614Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0376969Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0377280Z res = mod(**inputs) 2025-12-04T09:40:11.0377625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0377985Z outputs = self.fnet( 2025-12-04T09:40:11.0378330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0378699Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0379059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0379441Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0379797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0380154Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0380520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0380912Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0381305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0381680Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0382046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0382537Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0382686Z 2025-12-04T09:40:11.0382797Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0383139Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0383454Z res = mod(**inputs) 2025-12-04T09:40:11.0383789Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0384146Z outputs = self.fnet( 2025-12-04T09:40:11.0384477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0384839Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0385200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0385568Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0385920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0386267Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0386669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0387067Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0387453Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0387822Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0388187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0388609Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0388762Z 2025-12-04T09:40:11.0388842Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0389073Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0389408Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0389714Z res = mod(**inputs) 2025-12-04T09:40:11.0390056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0390425Z outputs = self.fnet( 2025-12-04T09:40:11.0390753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 511, in forward 2025-12-04T09:40:11.0391124Z embedding_output = self.embeddings( 2025-12-04T09:40:11.0391489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 141, in forward 2025-12-04T09:40:11.0391857Z embeddings = self.projection(embeddings) 2025-12-04T09:40:11.0391999Z 2025-12-04T09:40:11.0392085Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0392312Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0392649Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0392948Z res = mod(**inputs) 2025-12-04T09:40:11.0393286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0393639Z outputs = self.fnet( 2025-12-04T09:40:11.0393967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0394330Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0394687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0395060Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0395401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0395744Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0396109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0396494Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0396872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0397244Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0397609Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0397985Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0398138Z 2025-12-04T09:40:11.0398236Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0398571Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0398874Z res = mod(**inputs) 2025-12-04T09:40:11.0399201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0399576Z outputs = self.fnet( 2025-12-04T09:40:11.0399940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0400299Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0400644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0401018Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0401393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0401811Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0402194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0402610Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0403025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0403425Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0403823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0404217Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0404365Z 2025-12-04T09:40:11.0404474Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0404826Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0405164Z res = mod(**inputs) 2025-12-04T09:40:11.0405530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0405914Z outputs = self.fnet( 2025-12-04T09:40:11.0406279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0406676Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0407067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0407465Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0407841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0408217Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0408608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0409035Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0409445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0409848Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0410243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0410667Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0410823Z 2025-12-04T09:40:11.0410942Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0411309Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0411636Z res = mod(**inputs) 2025-12-04T09:40:11.0412006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0412395Z outputs = self.fnet( 2025-12-04T09:40:11.0412787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0413150Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0413509Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0413949Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0414292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0414635Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0415001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0415390Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0415771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0416137Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0416499Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0416875Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0417030Z 2025-12-04T09:40:11.0417110Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0417337Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0417672Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0417968Z res = mod(**inputs) 2025-12-04T09:40:11.0418299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0418650Z outputs = self.fnet( 2025-12-04T09:40:11.0418974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0419334Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0419687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0420058Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0420401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0420743Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0421104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 251, in forward 2025-12-04T09:40:11.0421468Z layer_output = apply_chunking_to_forward( 2025-12-04T09:40:11.0421859Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 257, in apply_chunking_to_forward 2025-12-04T09:40:11.0422240Z return forward_fn(*input_tensors) 2025-12-04T09:40:11.0422630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 260, in feed_forward_chunk 2025-12-04T09:40:11.0423054Z intermediate_output = self.intermediate(fourier_output) 2025-12-04T09:40:11.0423459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 219, in forward 2025-12-04T09:40:11.0423856Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-12-04T09:40:11.0424219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:40:11.0424649Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:40:11.0424878Z 2025-12-04T09:40:11.0424955Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0425161Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0425380Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0425722Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0426026Z res = mod(**inputs) 2025-12-04T09:40:11.0426363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0426775Z outputs = self.fnet( 2025-12-04T09:40:11.0427114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0427472Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0427823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0428209Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0428555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0428896Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0429251Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0429634Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0430020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0430595Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0430962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0431351Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0431503Z 2025-12-04T09:40:11.0431612Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0431952Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0432262Z res = mod(**inputs) 2025-12-04T09:40:11.0432601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0432959Z outputs = self.fnet( 2025-12-04T09:40:11.0433292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0433658Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0434017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0434384Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0434736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0435080Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0435445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0435825Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0436208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0436576Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0436944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0437326Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0437482Z 2025-12-04T09:40:11.0437583Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0437926Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0438227Z res = mod(**inputs) 2025-12-04T09:40:11.0438565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0438922Z outputs = self.fnet( 2025-12-04T09:40:11.0439259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0439611Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0440091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0440476Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0440823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0441171Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0441614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0442030Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0442425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0442825Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0443189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0443578Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0443724Z 2025-12-04T09:40:11.0443822Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0444178Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0444500Z res = mod(**inputs) 2025-12-04T09:40:11.0444848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0445224Z outputs = self.fnet( 2025-12-04T09:40:11.0445577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0445954Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0446323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0446716Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0447086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0447447Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0447827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0448231Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0448635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0449014Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0449400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0449805Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0449960Z 2025-12-04T09:40:11.0450048Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0450283Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0450645Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0450966Z res = mod(**inputs) 2025-12-04T09:40:11.0451312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0451686Z outputs = self.fnet( 2025-12-04T09:40:11.0452034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0452421Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0452770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0453142Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0453489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0453884Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0454242Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 251, in forward 2025-12-04T09:40:11.0454614Z layer_output = apply_chunking_to_forward( 2025-12-04T09:40:11.0454999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 257, in apply_chunking_to_forward 2025-12-04T09:40:11.0455387Z return forward_fn(*input_tensors) 2025-12-04T09:40:11.0455779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 260, in feed_forward_chunk 2025-12-04T09:40:11.0456208Z intermediate_output = self.intermediate(fourier_output) 2025-12-04T09:40:11.0456605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 219, in forward 2025-12-04T09:40:11.0456999Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-12-04T09:40:11.0457363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:40:11.0457794Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:40:11.0458014Z 2025-12-04T09:40:11.0458100Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0458300Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0458527Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0458867Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0459165Z res = mod(**inputs) 2025-12-04T09:40:11.0459512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0459864Z outputs = self.fnet( 2025-12-04T09:40:11.0460205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0460558Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0460912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0461282Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0461626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0461970Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0462333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0462713Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0463085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0463458Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0463819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0464204Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0464349Z 2025-12-04T09:40:11.0464448Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0464785Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0465091Z res = mod(**inputs) 2025-12-04T09:40:11.0465416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0465770Z outputs = self.fnet( 2025-12-04T09:40:11.0466103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0466482Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0466858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0467234Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0467582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0467920Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0468311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0468693Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0469072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0469429Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0469794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0470180Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0470327Z 2025-12-04T09:40:11.0470432Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0470764Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0471072Z res = mod(**inputs) 2025-12-04T09:40:11.0471405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0471754Z outputs = self.fnet( 2025-12-04T09:40:11.0472090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0472454Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0472810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0473181Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0473529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0473873Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0474229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0474611Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0474988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0475356Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0475712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0476097Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0476245Z 2025-12-04T09:40:11.0476355Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0476696Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0476999Z res = mod(**inputs) 2025-12-04T09:40:11.0477336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0477693Z outputs = self.fnet( 2025-12-04T09:40:11.0478022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0478386Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0478742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0479115Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0479457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0479848Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0480217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0480605Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0480995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0481389Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0481828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0482228Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0482386Z 2025-12-04T09:40:11.0482465Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0482713Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0483063Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0483367Z res = mod(**inputs) 2025-12-04T09:40:11.0483710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0484075Z outputs = self.fnet( 2025-12-04T09:40:11.0484412Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0484786Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0485150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0485531Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0486019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0486369Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0486749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 251, in forward 2025-12-04T09:40:11.0487134Z layer_output = apply_chunking_to_forward( 2025-12-04T09:40:11.0487523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 257, in apply_chunking_to_forward 2025-12-04T09:40:11.0487917Z return forward_fn(*input_tensors) 2025-12-04T09:40:11.0488318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 260, in feed_forward_chunk 2025-12-04T09:40:11.0488754Z intermediate_output = self.intermediate(fourier_output) 2025-12-04T09:40:11.0489165Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 219, in forward 2025-12-04T09:40:11.0489574Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-12-04T09:40:11.0489952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:40:11.0490385Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:40:11.0490615Z 2025-12-04T09:40:11.0490694Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0490902Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0491137Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0491488Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0491804Z res = mod(**inputs) 2025-12-04T09:40:11.0492154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0492511Z outputs = self.fnet( 2025-12-04T09:40:11.0492858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0493288Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0493655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0494029Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0494390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0494759Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0495126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0495518Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0495916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0496284Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0496644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0497030Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0497177Z 2025-12-04T09:40:11.0497285Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0497626Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0497922Z res = mod(**inputs) 2025-12-04T09:40:11.0498255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0498604Z outputs = self.fnet( 2025-12-04T09:40:11.0498927Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0499289Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0499642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0500010Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0500347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0500687Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0501050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0501425Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0501804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0502168Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0502529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0502907Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0503061Z 2025-12-04T09:40:11.0503161Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0503500Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0503802Z res = mod(**inputs) 2025-12-04T09:40:11.0504128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0504481Z outputs = self.fnet( 2025-12-04T09:40:11.0504814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0505167Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0505521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0505890Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0506285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0506622Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0506987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0507368Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0507763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0508138Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0508508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0508900Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0509046Z 2025-12-04T09:40:11.0509145Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0509490Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0509796Z res = mod(**inputs) 2025-12-04T09:40:11.0510132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0510482Z outputs = self.fnet( 2025-12-04T09:40:11.0510821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0511189Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0511540Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0511920Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0512273Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0512619Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0512982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0513371Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0513755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0514127Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0514489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0514881Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0515030Z 2025-12-04T09:40:11.0515113Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0515336Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0515678Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0515988Z res = mod(**inputs) 2025-12-04T09:40:11.0516327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0516677Z outputs = self.fnet( 2025-12-04T09:40:11.0517014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0517379Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0517731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0518110Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0518459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0518804Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0519194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 251, in forward 2025-12-04T09:40:11.0519597Z layer_output = apply_chunking_to_forward( 2025-12-04T09:40:11.0519984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 257, in apply_chunking_to_forward 2025-12-04T09:40:11.0520355Z return forward_fn(*input_tensors) 2025-12-04T09:40:11.0520744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 260, in feed_forward_chunk 2025-12-04T09:40:11.0521187Z intermediate_output = self.intermediate(fourier_output) 2025-12-04T09:40:11.0521687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 219, in forward 2025-12-04T09:40:11.0522126Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-12-04T09:40:11.0522535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:40:11.0522992Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:40:11.0523222Z 2025-12-04T09:40:11.0523320Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0523521Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0523755Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0524103Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0524407Z res = mod(**inputs) 2025-12-04T09:40:11.0524750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0525111Z outputs = self.fnet( 2025-12-04T09:40:11.0525457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0525806Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0526166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0526538Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0526887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0527222Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0527588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0527969Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0528343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0528710Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0529074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0529465Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0529612Z 2025-12-04T09:40:11.0529713Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0530052Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0530631Z res = mod(**inputs) 2025-12-04T09:40:11.0530965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0531321Z outputs = self.fnet( 2025-12-04T09:40:11.0531660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0532021Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0532369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0532796Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0533203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0533557Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0533956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0534368Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0534748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0535144Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0535508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0535894Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0536042Z 2025-12-04T09:40:11.0536152Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0536485Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0536788Z res = mod(**inputs) 2025-12-04T09:40:11.0537122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0537475Z outputs = self.fnet( 2025-12-04T09:40:11.0537804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0538163Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0538520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0538882Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0539230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0539573Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0539935Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0540307Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0540685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0541051Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0541404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0541788Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0541939Z 2025-12-04T09:40:11.0542037Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0542376Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0542675Z res = mod(**inputs) 2025-12-04T09:40:11.0543007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0543359Z outputs = self.fnet( 2025-12-04T09:40:11.0543693Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0544046Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0544396Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0544763Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0545103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0545443Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0545839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0546242Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0546613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0546979Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0547340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0547732Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0547887Z 2025-12-04T09:40:11.0547968Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0548199Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0548539Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0548838Z res = mod(**inputs) 2025-12-04T09:40:11.0549181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0549540Z outputs = self.fnet( 2025-12-04T09:40:11.0549871Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0550232Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0550588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0550965Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0551310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0551656Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0552025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 251, in forward 2025-12-04T09:40:11.0552400Z layer_output = apply_chunking_to_forward( 2025-12-04T09:40:11.0552783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 257, in apply_chunking_to_forward 2025-12-04T09:40:11.0553164Z return forward_fn(*input_tensors) 2025-12-04T09:40:11.0553555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 260, in feed_forward_chunk 2025-12-04T09:40:11.0553979Z intermediate_output = self.intermediate(fourier_output) 2025-12-04T09:40:11.0554379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 219, in forward 2025-12-04T09:40:11.0554776Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-12-04T09:40:11.0555141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:40:11.0555564Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:40:11.0555797Z 2025-12-04T09:40:11.0555879Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0556089Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0556328Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0556668Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0556977Z res = mod(**inputs) 2025-12-04T09:40:11.0557318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0557673Z outputs = self.fnet( 2025-12-04T09:40:11.0558013Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0558376Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0558736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0559153Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0559505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0559846Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0560205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0560650Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0561042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0561420Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0561853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0562256Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0562419Z 2025-12-04T09:40:11.0562530Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0562892Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0563206Z res = mod(**inputs) 2025-12-04T09:40:11.0563567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0563946Z outputs = self.fnet( 2025-12-04T09:40:11.0564349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0564737Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0565104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0565540Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0565896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0566247Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0566623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0567016Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0567399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0567776Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0568151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0568538Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0568696Z 2025-12-04T09:40:11.0568798Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0569150Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0569465Z res = mod(**inputs) 2025-12-04T09:40:11.0569801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0570165Z outputs = self.fnet( 2025-12-04T09:40:11.0570507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0570871Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0571236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0571617Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0571976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0572318Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0572756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0573152Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0573545Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0573915Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0574304Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0574700Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0574849Z 2025-12-04T09:40:11.0574950Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0575301Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0575621Z res = mod(**inputs) 2025-12-04T09:40:11.0575962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0576315Z outputs = self.fnet( 2025-12-04T09:40:11.0576647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0577005Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0577348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0577722Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0578068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0578409Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0578763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0579148Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0579528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0579898Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0580253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0580639Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0580785Z 2025-12-04T09:40:11.0580872Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0581094Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0581433Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0581737Z res = mod(**inputs) 2025-12-04T09:40:11.0582069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0582419Z outputs = self.fnet( 2025-12-04T09:40:11.0582751Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0583106Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0583453Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0583823Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0584169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0584509Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0584864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 251, in forward 2025-12-04T09:40:11.0585235Z layer_output = apply_chunking_to_forward( 2025-12-04T09:40:11.0585669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 257, in apply_chunking_to_forward 2025-12-04T09:40:11.0586047Z return forward_fn(*input_tensors) 2025-12-04T09:40:11.0586427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 260, in feed_forward_chunk 2025-12-04T09:40:11.0586854Z intermediate_output = self.intermediate(fourier_output) 2025-12-04T09:40:11.0587271Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 219, in forward 2025-12-04T09:40:11.0587662Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-12-04T09:40:11.0588025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:40:11.0588458Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:40:11.0588677Z 2025-12-04T09:40:11.0588765Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0588970Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0589203Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0589550Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0589864Z res = mod(**inputs) 2025-12-04T09:40:11.0590196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0590556Z outputs = self.fnet( 2025-12-04T09:40:11.0590896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0591249Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0591606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0591992Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0592343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0592679Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0593045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0593433Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0593807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0594174Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0594540Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0594925Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0595072Z 2025-12-04T09:40:11.0595173Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0595521Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0595829Z res = mod(**inputs) 2025-12-04T09:40:11.0596167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0596515Z outputs = self.fnet( 2025-12-04T09:40:11.0596854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0597217Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0597565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0597940Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0598287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0598649Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0599055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0599444Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0599831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0600206Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0600568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0600950Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0601095Z 2025-12-04T09:40:11.0601199Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0601530Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0601913Z res = mod(**inputs) 2025-12-04T09:40:11.0602257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0602624Z outputs = self.fnet( 2025-12-04T09:40:11.0602962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0603331Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0603707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0604073Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0604423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0604769Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0605147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0605540Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0605931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0606309Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0606678Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0607076Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0607235Z 2025-12-04T09:40:11.0607337Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0607685Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0607992Z res = mod(**inputs) 2025-12-04T09:40:11.0608341Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0608708Z outputs = self.fnet( 2025-12-04T09:40:11.0609050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0609409Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0609770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0610157Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0610507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0610859Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0611232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0611622Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0612060Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0612438Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0612812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0613213Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0613382Z 2025-12-04T09:40:11.0613462Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0613697Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0614045Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0614350Z res = mod(**inputs) 2025-12-04T09:40:11.0614695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0615060Z outputs = self.fnet( 2025-12-04T09:40:11.0615409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0615771Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0616134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0616538Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0616879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0617219Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0617583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 251, in forward 2025-12-04T09:40:11.0617952Z layer_output = apply_chunking_to_forward( 2025-12-04T09:40:11.0618329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 257, in apply_chunking_to_forward 2025-12-04T09:40:11.0618709Z return forward_fn(*input_tensors) 2025-12-04T09:40:11.0619097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 260, in feed_forward_chunk 2025-12-04T09:40:11.0619525Z intermediate_output = self.intermediate(fourier_output) 2025-12-04T09:40:11.0619920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 219, in forward 2025-12-04T09:40:11.0620317Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-12-04T09:40:11.0620681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:40:11.0621103Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:40:11.0621328Z 2025-12-04T09:40:11.0621405Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0621614Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0621851Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0622189Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0622496Z res = mod(**inputs) 2025-12-04T09:40:11.0622837Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0623182Z outputs = self.fnet( 2025-12-04T09:40:11.0623521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0623878Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0624232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0624598Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0624947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0625991Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0626391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0626767Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0627148Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0627531Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0627888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0628277Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0628431Z 2025-12-04T09:40:11.0628529Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0628865Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0629165Z res = mod(**inputs) 2025-12-04T09:40:11.0629498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0629853Z outputs = self.fnet( 2025-12-04T09:40:11.0630327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0630700Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0631059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0631435Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0631786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0632132Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0632505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0632889Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0633261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0633628Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0633994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0634374Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0634527Z 2025-12-04T09:40:11.0634627Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0634967Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0635275Z res = mod(**inputs) 2025-12-04T09:40:11.0635606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0635967Z outputs = self.fnet( 2025-12-04T09:40:11.0636305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0636665Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0637014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0637388Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0637735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0638069Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0638439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0638826Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0639295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0639656Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0640018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0640432Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0640580Z 2025-12-04T09:40:11.0640680Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0641020Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0641369Z res = mod(**inputs) 2025-12-04T09:40:11.0641777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0642150Z outputs = self.fnet( 2025-12-04T09:40:11.0642508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0642887Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0643277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0643666Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0644037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0644401Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0644782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0645188Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0645587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0645977Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0646352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0646761Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0646919Z 2025-12-04T09:40:11.0647007Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0647240Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0647600Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0647920Z res = mod(**inputs) 2025-12-04T09:40:11.0648272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0648636Z outputs = self.fnet( 2025-12-04T09:40:11.0648989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0649372Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0649745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0650127Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0650487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0650843Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0651219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 251, in forward 2025-12-04T09:40:11.0651608Z layer_output = apply_chunking_to_forward( 2025-12-04T09:40:11.0651990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 257, in apply_chunking_to_forward 2025-12-04T09:40:11.0652364Z return forward_fn(*input_tensors) 2025-12-04T09:40:11.0652801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 260, in feed_forward_chunk 2025-12-04T09:40:11.0653229Z intermediate_output = self.intermediate(fourier_output) 2025-12-04T09:40:11.0653624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 219, in forward 2025-12-04T09:40:11.0654017Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-12-04T09:40:11.0654388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:40:11.0654815Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:40:11.0655029Z 2025-12-04T09:40:11.0655113Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0655308Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0655535Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0655880Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0656186Z res = mod(**inputs) 2025-12-04T09:40:11.0656519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0656875Z outputs = self.fnet( 2025-12-04T09:40:11.0657213Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0657568Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0657925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0658297Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0658644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0658976Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0659344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0659733Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0660103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0660473Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0660838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0661228Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0661374Z 2025-12-04T09:40:11.0661475Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0661813Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0662116Z res = mod(**inputs) 2025-12-04T09:40:11.0662454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0662800Z outputs = self.fnet( 2025-12-04T09:40:11.0663135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0663501Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0663850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0664225Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0664572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0664913Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0665266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0665698Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0666078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0666448Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0666811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0667225Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0667372Z 2025-12-04T09:40:11.0667478Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0667814Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0668117Z res = mod(**inputs) 2025-12-04T09:40:11.0668449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0668804Z outputs = self.fnet( 2025-12-04T09:40:11.0669137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0669498Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0669855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0670222Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0670571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0670912Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0671274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0671646Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0672028Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0672397Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0672760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0673138Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0673290Z 2025-12-04T09:40:11.0673391Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0673729Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0674024Z res = mod(**inputs) 2025-12-04T09:40:11.0674360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0674715Z outputs = self.fnet( 2025-12-04T09:40:11.0675046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0675401Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0675757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0676126Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0676465Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0676816Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0677180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0677561Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0677935Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0678299Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0678719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0679105Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0679251Z 2025-12-04T09:40:11.0679328Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0679556Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0679915Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0680214Z res = mod(**inputs) 2025-12-04T09:40:11.0680550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0680906Z outputs = self.fnet( 2025-12-04T09:40:11.0681139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0681218Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0681459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0681652Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0681889Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0681968Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0682219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 251, in forward 2025-12-04T09:40:11.0682305Z layer_output = apply_chunking_to_forward( 2025-12-04T09:40:11.0682560Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 257, in apply_chunking_to_forward 2025-12-04T09:40:11.0682648Z return forward_fn(*input_tensors) 2025-12-04T09:40:11.0682916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 260, in feed_forward_chunk 2025-12-04T09:40:11.0683034Z intermediate_output = self.intermediate(fourier_output) 2025-12-04T09:40:11.0683282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 219, in forward 2025-12-04T09:40:11.0683391Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-12-04T09:40:11.0683607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:40:11.0683783Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:40:11.0683786Z 2025-12-04T09:40:11.0683865Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0683951Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0684054Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0684260Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0684325Z res = mod(**inputs) 2025-12-04T09:40:11.0684570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0684646Z outputs = self.fnet( 2025-12-04T09:40:11.0684882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0684955Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0685198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0685282Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0685509Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0685587Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0685858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0685978Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0686218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0686303Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0686543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0686659Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0686663Z 2025-12-04T09:40:11.0686773Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0686964Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0687028Z res = mod(**inputs) 2025-12-04T09:40:11.0687274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0687346Z outputs = self.fnet( 2025-12-04T09:40:11.0687588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0687661Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0687897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0687990Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0688204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0688281Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0688526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0688620Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0688867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0688946Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0689187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0689291Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0689296Z 2025-12-04T09:40:11.0689396Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0689593Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0689656Z res = mod(**inputs) 2025-12-04T09:40:11.0689893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0689967Z outputs = self.fnet( 2025-12-04T09:40:11.0690205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0690277Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0690520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0690603Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0690823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0690902Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0691140Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0691239Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0691477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0691578Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0691840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0691940Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0691943Z 2025-12-04T09:40:11.0692049Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0692237Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0692317Z res = mod(**inputs) 2025-12-04T09:40:11.0692563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0692627Z outputs = self.fnet( 2025-12-04T09:40:11.0692870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0692941Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0693182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0693272Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0693483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0693568Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0693812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0693905Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0694146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0694224Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0694459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0694567Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0694570Z 2025-12-04T09:40:11.0694648Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0694754Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0694944Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0695008Z res = mod(**inputs) 2025-12-04T09:40:11.0695251Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0695317Z outputs = self.fnet( 2025-12-04T09:40:11.0695552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0695629Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0695864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0695956Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0696168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0696243Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0696487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 251, in forward 2025-12-04T09:40:11.0696571Z layer_output = apply_chunking_to_forward( 2025-12-04T09:40:11.0696840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 257, in apply_chunking_to_forward 2025-12-04T09:40:11.0696915Z return forward_fn(*input_tensors) 2025-12-04T09:40:11.0697172Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 260, in feed_forward_chunk 2025-12-04T09:40:11.0697286Z intermediate_output = self.intermediate(fourier_output) 2025-12-04T09:40:11.0697568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 219, in forward 2025-12-04T09:40:11.0697672Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-12-04T09:40:11.0697880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:40:11.0698063Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:40:11.0698067Z 2025-12-04T09:40:11.0698149Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0698223Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0698320Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0698514Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0698575Z res = mod(**inputs) 2025-12-04T09:40:11.0698819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0698881Z outputs = self.fnet( 2025-12-04T09:40:11.0699124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0699200Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0699431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0699514Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0699734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0699808Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0700045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0700137Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0700370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0700454Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0700686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0700784Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0700794Z 2025-12-04T09:40:11.0700890Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0701073Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0701140Z res = mod(**inputs) 2025-12-04T09:40:11.0701370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0701433Z outputs = self.fnet( 2025-12-04T09:40:11.0701673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0701742Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0701977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0702057Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0702272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0702354Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0702586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0702677Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0702919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0703040Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0703278Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0703373Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0703376Z 2025-12-04T09:40:11.0703474Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0703683Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0703745Z res = mod(**inputs) 2025-12-04T09:40:11.0703985Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0704048Z outputs = self.fnet( 2025-12-04T09:40:11.0704281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0704359Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0704595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0704674Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0704889Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0704965Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0705204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0705294Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0705527Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0705609Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0705843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0705948Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0705951Z 2025-12-04T09:40:11.0706048Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0706235Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0706304Z res = mod(**inputs) 2025-12-04T09:40:11.0706537Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0706600Z outputs = self.fnet( 2025-12-04T09:40:11.0706837Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0706906Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0707145Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0707229Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0707440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0707521Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0707754Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0707852Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0708085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0708160Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0708398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0708493Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0708519Z 2025-12-04T09:40:11.0708628Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0708738Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0708926Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0708992Z res = mod(**inputs) 2025-12-04T09:40:11.0709221Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0709300Z outputs = self.fnet( 2025-12-04T09:40:11.0709538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0709607Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0709839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0709929Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0710141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0710223Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0710454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 251, in forward 2025-12-04T09:40:11.0710535Z layer_output = apply_chunking_to_forward( 2025-12-04T09:40:11.0710793Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 257, in apply_chunking_to_forward 2025-12-04T09:40:11.0710867Z return forward_fn(*input_tensors) 2025-12-04T09:40:11.0711131Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 260, in feed_forward_chunk 2025-12-04T09:40:11.0711240Z intermediate_output = self.intermediate(fourier_output) 2025-12-04T09:40:11.0711476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 219, in forward 2025-12-04T09:40:11.0711589Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-12-04T09:40:11.0711793Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:40:11.0711962Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:40:11.0711974Z 2025-12-04T09:40:11.0712050Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0712124Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0712228Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0712417Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0712478Z res = mod(**inputs) 2025-12-04T09:40:11.0712719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0712789Z outputs = self.fnet( 2025-12-04T09:40:11.0713028Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0713098Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0713336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0713423Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0713631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0713707Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0713946Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0714038Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0714323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0714402Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0714632Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0714736Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0714755Z 2025-12-04T09:40:11.0714854Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0715046Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0715107Z res = mod(**inputs) 2025-12-04T09:40:11.0715337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0715407Z outputs = self.fnet( 2025-12-04T09:40:11.0715636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0715710Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0715948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0716028Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0716243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0716318Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0716550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0716647Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0716877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0716952Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0717193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0717288Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0717291Z 2025-12-04T09:40:11.0717398Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0717581Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0717644Z res = mod(**inputs) 2025-12-04T09:40:11.0717888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0717954Z outputs = self.fnet( 2025-12-04T09:40:11.0718202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0718274Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0718513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0718606Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0718816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0718892Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0719143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0719239Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0719483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0719561Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0719799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0719961Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0719965Z 2025-12-04T09:40:11.0720064Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0720262Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0720324Z res = mod(**inputs) 2025-12-04T09:40:11.0720562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0720650Z outputs = self.fnet( 2025-12-04T09:40:11.0720897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0720965Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0721219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0721300Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0721533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0721682Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0721932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 248, in forward 2025-12-04T09:40:11.0722033Z self_fourier_outputs = self.fourier(hidden_states) 2025-12-04T09:40:11.0722279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 201, in forward 2025-12-04T09:40:11.0722365Z self_outputs = self.self(hidden_states) 2025-12-04T09:40:11.0722606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 180, in forward 2025-12-04T09:40:11.0722703Z outputs = self.fourier_transform(hidden_states).real 2025-12-04T09:40:11.0722707Z 2025-12-04T09:40:11.0722793Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0722899Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0723091Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0723164Z res = mod(**inputs) 2025-12-04T09:40:11.0723409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 670, in forward 2025-12-04T09:40:11.0723482Z outputs = self.fnet( 2025-12-04T09:40:11.0723724Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 517, in forward 2025-12-04T09:40:11.0723796Z encoder_outputs = self.encoder( 2025-12-04T09:40:11.0724048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 279, in forward 2025-12-04T09:40:11.0724131Z layer_outputs = layer_module(hidden_states) 2025-12-04T09:40:11.0724348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:40:11.0724435Z return super().__call__(*args, **kwargs) 2025-12-04T09:40:11.0724676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 251, in forward 2025-12-04T09:40:11.0724768Z layer_output = apply_chunking_to_forward( 2025-12-04T09:40:11.0725025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 257, in apply_chunking_to_forward 2025-12-04T09:40:11.0725102Z return forward_fn(*input_tensors) 2025-12-04T09:40:11.0725379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 260, in feed_forward_chunk 2025-12-04T09:40:11.0725490Z intermediate_output = self.intermediate(fourier_output) 2025-12-04T09:40:11.0725742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 219, in forward 2025-12-04T09:40:11.0725901Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-12-04T09:40:11.0726110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 62, in forward 2025-12-04T09:40:11.0726287Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-12-04T09:40:11.0726290Z 2025-12-04T09:40:11.0726368Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0726460Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0726541Z cudagraph partition due to non gpu ops 2025-12-04T09:40:11.0726643Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:11.0726840Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:11.0726902Z res = mod(**inputs) 2025-12-04T09:40:11.0727143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/fnet/modeling_fnet.py", line 685, in forward 2025-12-04T09:40:11.0727335Z masked_lm_loss = loss_fct(prediction_scores.view(-1, self.config.vocab_size), labels.view(-1)) 2025-12-04T09:40:11.0727338Z 2025-12-04T09:40:20.5727408Z Compilation time (from dynamo_timed): 14.235906889 2025-12-04T09:40:20.5780938Z pass 2025-12-04T09:40:20.5781428Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:40:20.5785052Z TIMING: _recursive_pre_grad_passes:0.02014 _recursive_joint_graph_passes:0.20094 _recursive_post_grad_passes:0.05393 async_compile.wait:0.76788 code_gen:8.98708 inductor_compile:10.36738 backend_compile:12.43863 gc:0.00107 entire_frame_compile:14.23591 total_wall_time:14.23591 2025-12-04T09:40:20.5786023Z STATS: call_* op count: 232 | FakeTensorMode.__torch_dispatch__:11067 | FakeTensor.__torch_dispatch__:2950 | ProxyTorchDispatchMode.__torch_dispatch__:1323 2025-12-04T09:40:20.5790403Z Dynamo produced 1 graphs covering 232 ops with 0 graph breaks (0 unique) 2025-12-04T09:40:22.7635998Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:40:22.7636909Z import pynvml # type: ignore[import] 2025-12-04T09:40:26.0009525Z 2025-12-04T09:40:27.0203599Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:40:27.0204056Z loading model: 0it [00:01, ?it/s] 2025-12-04T09:40:27.0204324Z cpu eval LayoutLMForMaskedLM 2025-12-04T09:40:27.5452491Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:40:27.6807615Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:40:27.8155313Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:40:38.0616493Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0616912Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0617177Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0617419Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0617655Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0617892Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0618116Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0618362Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0618594Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0618816Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0619049Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0619273Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0619505Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0619727Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0619952Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0620196Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0620837Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0621194Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0621432Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0621659Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0621898Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0622124Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0622344Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0622713Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0622946Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0623291Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0623521Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0623754Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0624029Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0624251Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0624488Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0624728Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0624961Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0625197Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0625426Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0625654Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0625880Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0626114Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0626348Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0626572Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0626808Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0627020Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0627229Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0627442Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0627665Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0627884Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0628111Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0628338Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0628564Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0628827Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0629053Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0629278Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0629499Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0629722Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0629934Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0630304Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0630531Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0630747Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0630961Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0631175Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0631401Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0631627Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0631847Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0632072Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0632299Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0632522Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0632736Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0632948Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0633160Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0633369Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0633579Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0633792Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0633997Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0634210Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0634427Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0634631Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0634905Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0635235Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0635450Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0635665Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0635875Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0636080Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0636294Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0636539Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0636758Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0636970Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0637181Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0637394Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0637613Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0637832Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0638053Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0638268Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0638491Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0638712Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0638924Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0639143Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0639361Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0639574Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0639793Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0640011Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0640231Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0640445Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0640666Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0640886Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0641097Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0641316Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0641711Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0641940Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0642162Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0642386Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0642605Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0642834Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0643064Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0643276Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0643482Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0643704Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0643926Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0644145Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0644368Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0644592Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0644806Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0645029Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0645253Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0645469Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0645688Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0645909Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0646138Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0646351Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0646576Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0646796Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0647010Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0647231Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0647455Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0647672Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0647893Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0648115Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0648372Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0648624Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0648848Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0649068Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0649281Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0649501Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0649721Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0649968Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0650193Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0650415Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0650638Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0650855Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0651078Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0651297Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0651518Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0651745Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0651973Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0652187Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0652409Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0652630Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0652845Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0653069Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0653297Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0653522Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0653738Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0653974Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0654199Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0654415Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0654638Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0654860Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0655078Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0655306Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0655533Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0655756Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0655973Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0656196Z cudagraph partition due to non gpu ops 2025-12-04T09:40:38.0656459Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:40:38.0656876Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:40:38.0657230Z res = mod(**inputs) 2025-12-04T09:40:38.0657614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:40:38.0657991Z output = func(self, *args, **kwargs) 2025-12-04T09:40:38.0658422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/layoutlm/modeling_layoutlm.py", line 740, in forward 2025-12-04T09:40:38.0658848Z masked_lm_loss = loss_fct( 2025-12-04T09:40:38.0658983Z 2025-12-04T09:40:47.4167517Z Compilation time (from dynamo_timed): 18.991471209 2025-12-04T09:40:47.4201199Z pass 2025-12-04T09:40:47.4201717Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:40:47.4202793Z TIMING: _recursive_pre_grad_passes:0.03757 _recursive_joint_graph_passes:0.46077 _recursive_post_grad_passes:0.06572 async_compile.wait:0.64557 code_gen:8.96513 inductor_compile:11.02776 backend_compile:16.03462 gc:0.0006 entire_frame_compile:18.99147 total_wall_time:18.99147 2025-12-04T09:40:47.4204001Z STATS: call_* op count: 434 | FakeTensorMode.__torch_dispatch__:22389 | FakeTensor.__torch_dispatch__:3051 | ProxyTorchDispatchMode.__torch_dispatch__:4767 2025-12-04T09:40:47.4204551Z Dynamo produced 1 graphs covering 434 ops with 0 graph breaks (0 unique) 2025-12-04T09:40:49.8982416Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:40:49.8983276Z import pynvml # type: ignore[import] 2025-12-04T09:40:53.1668588Z 2025-12-04T09:40:58.6822116Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:40:58.6829676Z loading model: 0it [00:05, ?it/s] 2025-12-04T09:40:58.6831948Z cpu eval M2M100ForConditionalGeneration 2025-12-04T09:40:59.5010053Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:40:59.8994628Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:41:00.2642657Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:41:20.8091091Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8095766Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8100456Z res = mod(**inputs) 2025-12-04T09:41:20.8105578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8108368Z outputs = self.model( 2025-12-04T09:41:20.8113347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8113892Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8120027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 849, in forward 2025-12-04T09:41:20.8120690Z embed_pos = self.embed_positions(input_ids, inputs_embeds) 2025-12-04T09:41:20.8121254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 124, in decorate_context 2025-12-04T09:41:20.8127078Z return func(*args, **kwargs) 2025-12-04T09:41:20.8127622Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 149, in forward 2025-12-04T09:41:20.8128169Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length).to( 2025-12-04T09:41:20.8133865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 81, in create_position_ids_from_input_ids 2025-12-04T09:41:20.8139260Z mask = input_ids.ne(padding_idx).int() 2025-12-04T09:41:20.8139491Z 2025-12-04T09:41:20.8139586Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8139817Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8140024Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8140239Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8140450Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8140655Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8140870Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8141084Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8141292Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8141493Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8141702Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8141912Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8142150Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8142542Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8142899Z res = mod(**inputs) 2025-12-04T09:41:20.8143313Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8143699Z outputs = self.model( 2025-12-04T09:41:20.8144072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8144875Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8145261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 849, in forward 2025-12-04T09:41:20.8145686Z embed_pos = self.embed_positions(input_ids, inputs_embeds) 2025-12-04T09:41:20.8146102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 124, in decorate_context 2025-12-04T09:41:20.8146520Z return func(*args, **kwargs) 2025-12-04T09:41:20.8146903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 149, in forward 2025-12-04T09:41:20.8147435Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length).to( 2025-12-04T09:41:20.8148020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 82, in create_position_ids_from_input_ids 2025-12-04T09:41:20.8148584Z incremental_indices = (torch.cumsum(mask, dim=1).type_as(mask) + past_key_values_length) * mask 2025-12-04T09:41:20.8148821Z 2025-12-04T09:41:20.8148932Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8149300Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8149628Z res = mod(**inputs) 2025-12-04T09:41:20.8149994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8150387Z outputs = self.model( 2025-12-04T09:41:20.8150763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8151157Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8151521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 849, in forward 2025-12-04T09:41:20.8151937Z embed_pos = self.embed_positions(input_ids, inputs_embeds) 2025-12-04T09:41:20.8152329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 124, in decorate_context 2025-12-04T09:41:20.8152692Z return func(*args, **kwargs) 2025-12-04T09:41:20.8153070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 149, in forward 2025-12-04T09:41:20.8153577Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length).to( 2025-12-04T09:41:20.8154147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 82, in create_position_ids_from_input_ids 2025-12-04T09:41:20.8154698Z incremental_indices = (torch.cumsum(mask, dim=1).type_as(mask) + past_key_values_length) * mask 2025-12-04T09:41:20.8154933Z 2025-12-04T09:41:20.8155016Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8155233Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8155465Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8155668Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8155878Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8156091Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8156296Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8156539Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8156916Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8157259Z res = mod(**inputs) 2025-12-04T09:41:20.8157619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8158016Z outputs = self.model( 2025-12-04T09:41:20.8158375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8158853Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8159241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8159650Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8160078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8160558Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8161001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 381, in forward 2025-12-04T09:41:20.8161451Z hidden_states, attn_weights = self.self_attn( 2025-12-04T09:41:20.8162016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8162433Z return func(*args, **kwargs) 2025-12-04T09:41:20.8162855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8163306Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8163794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8164321Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8164534Z 2025-12-04T09:41:20.8164622Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8164853Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8165074Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8165298Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8165521Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8165762Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8165988Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8166214Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8166431Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8166655Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8166877Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8167092Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8167316Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8167572Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8167960Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8168320Z res = mod(**inputs) 2025-12-04T09:41:20.8168724Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8169165Z outputs = self.model( 2025-12-04T09:41:20.8169565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8170014Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8170449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8170871Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8171257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8171639Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8172038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 381, in forward 2025-12-04T09:41:20.8172426Z hidden_states, attn_weights = self.self_attn( 2025-12-04T09:41:20.8172813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8173180Z return func(*args, **kwargs) 2025-12-04T09:41:20.8173786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8174212Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8174656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8175117Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8175309Z 2025-12-04T09:41:20.8175397Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8175601Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8175810Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8176014Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8176213Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8176417Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8176652Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8177010Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8177340Z res = mod(**inputs) 2025-12-04T09:41:20.8177709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8178095Z outputs = self.model( 2025-12-04T09:41:20.8178464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8178852Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8179235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8179611Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8179963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8180335Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8180741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 396, in forward 2025-12-04T09:41:20.8181138Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8181280Z 2025-12-04T09:41:20.8181357Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8181571Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8181786Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8181992Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8182212Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8182418Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8182615Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8182852Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8183228Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8183545Z res = mod(**inputs) 2025-12-04T09:41:20.8183916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8184301Z outputs = self.model( 2025-12-04T09:41:20.8184673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8185052Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8185431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8185816Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8186160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8186523Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8186913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 381, in forward 2025-12-04T09:41:20.8187348Z hidden_states, attn_weights = self.self_attn( 2025-12-04T09:41:20.8187798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8188173Z return func(*args, **kwargs) 2025-12-04T09:41:20.8188547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8188979Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8189414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8189888Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8190066Z 2025-12-04T09:41:20.8190151Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8190356Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8190567Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8190777Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8190984Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8191180Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8191383Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8191586Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8191783Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8191988Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8192193Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8192389Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8192592Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8192828Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8193181Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8193503Z res = mod(**inputs) 2025-12-04T09:41:20.8193875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8194258Z outputs = self.model( 2025-12-04T09:41:20.8194618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8195006Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8195384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8195760Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8196110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8196474Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8196864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 381, in forward 2025-12-04T09:41:20.8197259Z hidden_states, attn_weights = self.self_attn( 2025-12-04T09:41:20.8197667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8198029Z return func(*args, **kwargs) 2025-12-04T09:41:20.8198389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8198775Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8199207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8199666Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8199839Z 2025-12-04T09:41:20.8199925Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8200124Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8200324Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8200523Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8200775Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8200976Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8201207Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8201636Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8201964Z res = mod(**inputs) 2025-12-04T09:41:20.8202353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8202729Z outputs = self.model( 2025-12-04T09:41:20.8203090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8203472Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8203857Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8204260Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8204619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8204987Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8205382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 396, in forward 2025-12-04T09:41:20.8205763Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8205905Z 2025-12-04T09:41:20.8205981Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8206187Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8206384Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8206586Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8206790Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8206992Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8207185Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8207426Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8207780Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8208094Z res = mod(**inputs) 2025-12-04T09:41:20.8208466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8208863Z outputs = self.model( 2025-12-04T09:41:20.8209218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8209599Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8209974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8210350Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8210692Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8211042Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8211424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 381, in forward 2025-12-04T09:41:20.8211818Z hidden_states, attn_weights = self.self_attn( 2025-12-04T09:41:20.8212201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8212572Z return func(*args, **kwargs) 2025-12-04T09:41:20.8212938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8213333Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8213766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8214253Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8214489Z 2025-12-04T09:41:20.8214578Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8214779Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8214983Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8215183Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8215376Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8215600Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8215805Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8216003Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8216208Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8216414Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8216617Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8216811Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8217011Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8217239Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8217588Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8217903Z res = mod(**inputs) 2025-12-04T09:41:20.8218267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8218648Z outputs = self.model( 2025-12-04T09:41:20.8219009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8219391Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8219769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8220142Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8220486Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8220845Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8221231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 381, in forward 2025-12-04T09:41:20.8221621Z hidden_states, attn_weights = self.self_attn( 2025-12-04T09:41:20.8222012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8222386Z return func(*args, **kwargs) 2025-12-04T09:41:20.8222750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8223151Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8223585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8224054Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8224231Z 2025-12-04T09:41:20.8224312Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8224521Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8224725Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8224931Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8225135Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8225334Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8225565Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8225911Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8226227Z res = mod(**inputs) 2025-12-04T09:41:20.8226589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8226961Z outputs = self.model( 2025-12-04T09:41:20.8227366Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8227782Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8228156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8228522Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8228872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8229235Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8229606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 396, in forward 2025-12-04T09:41:20.8229991Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8230307Z 2025-12-04T09:41:20.8230392Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8230602Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8230802Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8231010Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8231223Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8231413Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8231612Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8231839Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8232187Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8232498Z res = mod(**inputs) 2025-12-04T09:41:20.8232857Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8233230Z outputs = self.model( 2025-12-04T09:41:20.8233593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8233961Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8234335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8234712Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8235050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8235407Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8235792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 381, in forward 2025-12-04T09:41:20.8236181Z hidden_states, attn_weights = self.self_attn( 2025-12-04T09:41:20.8236580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8236941Z return func(*args, **kwargs) 2025-12-04T09:41:20.8237297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8237680Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8238105Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8238567Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8238746Z 2025-12-04T09:41:20.8238830Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8239033Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8239235Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8239435Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8239631Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8239831Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8240030Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8240224Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8240427Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8240691Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8240953Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8241161Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8241363Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8241653Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8242001Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8242355Z res = mod(**inputs) 2025-12-04T09:41:20.8242723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8243104Z outputs = self.model( 2025-12-04T09:41:20.8243482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8243878Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8244259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8244628Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8244967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8245318Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8245691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 381, in forward 2025-12-04T09:41:20.8246085Z hidden_states, attn_weights = self.self_attn( 2025-12-04T09:41:20.8246468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8246832Z return func(*args, **kwargs) 2025-12-04T09:41:20.8247189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8247585Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8248019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8248476Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8248651Z 2025-12-04T09:41:20.8248727Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8248934Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8249138Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8249330Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8249528Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8249727Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8249946Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8250298Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8250612Z res = mod(**inputs) 2025-12-04T09:41:20.8250972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8251350Z outputs = self.model( 2025-12-04T09:41:20.8251717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8252110Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8252480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8252845Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8253187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8253546Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8253917Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 396, in forward 2025-12-04T09:41:20.8254355Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8254497Z 2025-12-04T09:41:20.8254574Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8254779Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8254974Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8255175Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8255373Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8255586Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8255790Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8256018Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8256362Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8256679Z res = mod(**inputs) 2025-12-04T09:41:20.8257037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8257414Z outputs = self.model( 2025-12-04T09:41:20.8257770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8258150Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8258522Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8258892Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8259236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8259591Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8259974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 381, in forward 2025-12-04T09:41:20.8260352Z hidden_states, attn_weights = self.self_attn( 2025-12-04T09:41:20.8260731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8261087Z return func(*args, **kwargs) 2025-12-04T09:41:20.8261445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8261825Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8262244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8262697Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8262867Z 2025-12-04T09:41:20.8262943Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8263146Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8263343Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8263540Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8263728Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8263929Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8264125Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8264315Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8264510Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8264706Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8264892Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8265093Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8265290Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8265510Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8265854Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8266161Z res = mod(**inputs) 2025-12-04T09:41:20.8266509Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8266893Z outputs = self.model( 2025-12-04T09:41:20.8267465Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8267848Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8268220Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8268588Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8268947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8269296Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8269665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 381, in forward 2025-12-04T09:41:20.8270051Z hidden_states, attn_weights = self.self_attn( 2025-12-04T09:41:20.8270427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8270788Z return func(*args, **kwargs) 2025-12-04T09:41:20.8271135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8271521Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8271939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8272385Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8272554Z 2025-12-04T09:41:20.8272629Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8272829Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8273028Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8273216Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8273411Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8273607Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8273822Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8274163Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8274467Z res = mod(**inputs) 2025-12-04T09:41:20.8274813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8275169Z outputs = self.model( 2025-12-04T09:41:20.8275515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8275878Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8276230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8276593Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8276928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8277273Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8277638Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 396, in forward 2025-12-04T09:41:20.8278007Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8278135Z 2025-12-04T09:41:20.8278216Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8278405Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8278601Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8278795Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8278987Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8279173Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8279365Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8279584Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8279973Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8280284Z res = mod(**inputs) 2025-12-04T09:41:20.8280634Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8281005Z outputs = self.model( 2025-12-04T09:41:20.8281355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8281856Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8282272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8282749Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8283111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8283489Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8283897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 381, in forward 2025-12-04T09:41:20.8284331Z hidden_states, attn_weights = self.self_attn( 2025-12-04T09:41:20.8284777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8285203Z return func(*args, **kwargs) 2025-12-04T09:41:20.8285617Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8286063Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8286549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8287063Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8287259Z 2025-12-04T09:41:20.8287348Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8287579Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8287804Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8288020Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8288241Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8288461Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8288685Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8288904Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8289124Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8289343Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8289556Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8289779Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8290001Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8290249Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8290592Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8290903Z res = mod(**inputs) 2025-12-04T09:41:20.8291252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8291608Z outputs = self.model( 2025-12-04T09:41:20.8291960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8292331Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8292689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8293056Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8293391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8293736Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8294177Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 381, in forward 2025-12-04T09:41:20.8294568Z hidden_states, attn_weights = self.self_attn( 2025-12-04T09:41:20.8294944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8295300Z return func(*args, **kwargs) 2025-12-04T09:41:20.8295674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8296074Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8296512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8296965Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8297145Z 2025-12-04T09:41:20.8297223Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8297435Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8297640Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8297835Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8298038Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8298238Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8298460Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8298809Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8299119Z res = mod(**inputs) 2025-12-04T09:41:20.8299466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8299835Z outputs = self.model( 2025-12-04T09:41:20.8300192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1258, in forward 2025-12-04T09:41:20.8300571Z encoder_outputs = self.encoder( 2025-12-04T09:41:20.8300933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 883, in forward 2025-12-04T09:41:20.8301305Z layer_outputs = encoder_layer( 2025-12-04T09:41:20.8301644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8301995Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8302371Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 396, in forward 2025-12-04T09:41:20.8302748Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8302882Z 2025-12-04T09:41:20.8302968Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8303169Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8303371Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8303572Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8303765Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8303967Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8304170Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8304396Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8304739Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8305051Z res = mod(**inputs) 2025-12-04T09:41:20.8305407Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8305770Z outputs = self.model( 2025-12-04T09:41:20.8306128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8306502Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8306870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8307282Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8307617Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8307960Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8308314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8308702Z return func(*args, **kwargs) 2025-12-04T09:41:20.8309063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 477, in forward 2025-12-04T09:41:20.8309456Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:20.8309834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8310189Z return func(*args, **kwargs) 2025-12-04T09:41:20.8310553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8310941Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8311357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8311811Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8311984Z 2025-12-04T09:41:20.8312070Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8312269Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8312471Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8312669Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8312865Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8313055Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8313251Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8313448Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8313638Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8313860Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8314203Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8314501Z res = mod(**inputs) 2025-12-04T09:41:20.8314851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8315216Z outputs = self.model( 2025-12-04T09:41:20.8315566Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8315931Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8316297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8316664Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8316995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8317343Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8317719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8318084Z return func(*args, **kwargs) 2025-12-04T09:41:20.8318484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 494, in forward 2025-12-04T09:41:20.8318898Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-12-04T09:41:20.8319305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8319666Z return func(*args, **kwargs) 2025-12-04T09:41:20.8320020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8320483Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8320932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8321401Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8321663Z 2025-12-04T09:41:20.8321773Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8321993Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8322207Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8322417Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8322628Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8322842Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8323036Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8323238Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8323442Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8323638Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8323841Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8324042Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8324243Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8324471Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8324830Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8325145Z res = mod(**inputs) 2025-12-04T09:41:20.8325502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8325877Z outputs = self.model( 2025-12-04T09:41:20.8326240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8326620Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8326994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8327376Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8327722Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8328069Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8328445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8328811Z return func(*args, **kwargs) 2025-12-04T09:41:20.8329176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 477, in forward 2025-12-04T09:41:20.8329571Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:20.8329962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8330482Z return func(*args, **kwargs) 2025-12-04T09:41:20.8330853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8331257Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8331693Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8332162Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8332340Z 2025-12-04T09:41:20.8332418Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8332627Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8332866Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8333221Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8333533Z res = mod(**inputs) 2025-12-04T09:41:20.8333996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8334374Z outputs = self.model( 2025-12-04T09:41:20.8334769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8335139Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8335545Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8335908Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8336236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8336581Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8336942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8337295Z return func(*args, **kwargs) 2025-12-04T09:41:20.8337654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 486, in forward 2025-12-04T09:41:20.8338026Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8338156Z 2025-12-04T09:41:20.8338240Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8338437Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8338639Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8338835Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8339027Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8339226Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8339426Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8339649Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8339989Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8340293Z res = mod(**inputs) 2025-12-04T09:41:20.8340639Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8340995Z outputs = self.model( 2025-12-04T09:41:20.8341343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8341713Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8342079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8342441Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8342782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8343141Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8343496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8343853Z return func(*args, **kwargs) 2025-12-04T09:41:20.8344219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 494, in forward 2025-12-04T09:41:20.8344618Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-12-04T09:41:20.8345000Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8345356Z return func(*args, **kwargs) 2025-12-04T09:41:20.8345713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8346096Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8346540Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8347016Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8347220Z 2025-12-04T09:41:20.8347310Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8347512Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8347719Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8347922Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8348116Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8348338Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8348536Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8348737Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8348937Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8349130Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8349321Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8349507Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8349701Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8349925Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8350265Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8350576Z res = mod(**inputs) 2025-12-04T09:41:20.8350925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8351300Z outputs = self.model( 2025-12-04T09:41:20.8351654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8352033Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8352417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8352779Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8353112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8353462Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8353827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8354179Z return func(*args, **kwargs) 2025-12-04T09:41:20.8354544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 477, in forward 2025-12-04T09:41:20.8354948Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:20.8355342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8355695Z return func(*args, **kwargs) 2025-12-04T09:41:20.8356058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8356459Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8356884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8357342Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8357524Z 2025-12-04T09:41:20.8357602Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8357811Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8358009Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8358211Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8358411Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8358604Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8358803Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8359000Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8359191Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8359421Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8359771Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8360146Z res = mod(**inputs) 2025-12-04T09:41:20.8360501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8360878Z outputs = self.model( 2025-12-04T09:41:20.8361239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8361709Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8362092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8362488Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8362842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8363203Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8363600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8363967Z return func(*args, **kwargs) 2025-12-04T09:41:20.8364329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 494, in forward 2025-12-04T09:41:20.8364745Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-12-04T09:41:20.8365154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8365521Z return func(*args, **kwargs) 2025-12-04T09:41:20.8365880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8366282Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8366717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8367184Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8367363Z 2025-12-04T09:41:20.8367442Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8367652Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8367897Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8368234Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8368541Z res = mod(**inputs) 2025-12-04T09:41:20.8368891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8369260Z outputs = self.model( 2025-12-04T09:41:20.8369603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8369971Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8370339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8370704Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8371037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8371383Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8371752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8372099Z return func(*args, **kwargs) 2025-12-04T09:41:20.8372458Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 503, in forward 2025-12-04T09:41:20.8372839Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8372967Z 2025-12-04T09:41:20.8373048Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8373291Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8373522Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8373723Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8373913Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8374111Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8374308Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8374497Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8374718Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8374910Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8375099Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8375325Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8375666Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8375972Z res = mod(**inputs) 2025-12-04T09:41:20.8376310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8376679Z outputs = self.model( 2025-12-04T09:41:20.8377028Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8377391Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8377754Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8378124Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8378455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8378791Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8379152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8379510Z return func(*args, **kwargs) 2025-12-04T09:41:20.8379865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 477, in forward 2025-12-04T09:41:20.8380253Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:20.8380633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8380991Z return func(*args, **kwargs) 2025-12-04T09:41:20.8381339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8381730Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8382147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8382598Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8382769Z 2025-12-04T09:41:20.8382845Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8383050Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8383249Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8383438Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8383635Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8383831Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8384025Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8384215Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8384412Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8384638Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8384981Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8385297Z res = mod(**inputs) 2025-12-04T09:41:20.8385656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8386026Z outputs = self.model( 2025-12-04T09:41:20.8386454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8386837Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8387215Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8387585Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8387949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8388304Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8388680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8389036Z return func(*args, **kwargs) 2025-12-04T09:41:20.8389403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 494, in forward 2025-12-04T09:41:20.8389824Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-12-04T09:41:20.8390219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8390587Z return func(*args, **kwargs) 2025-12-04T09:41:20.8390952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8391354Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8391781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8392244Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8392421Z 2025-12-04T09:41:20.8392505Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8392715Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8392913Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8393118Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8393320Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8393511Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8393739Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8394089Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8394398Z res = mod(**inputs) 2025-12-04T09:41:20.8394756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8395131Z outputs = self.model( 2025-12-04T09:41:20.8395490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8395863Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8396240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8396618Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8396952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8397304Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8397678Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8398041Z return func(*args, **kwargs) 2025-12-04T09:41:20.8398399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 512, in forward 2025-12-04T09:41:20.8398780Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8398912Z 2025-12-04T09:41:20.8398996Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8399192Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8399415Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8399664Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8399862Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8400053Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8400254Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8400483Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8400823Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8401157Z res = mod(**inputs) 2025-12-04T09:41:20.8401588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8401977Z outputs = self.model( 2025-12-04T09:41:20.8402338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8402731Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8403128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8403528Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8403913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8404274Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8404655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8405018Z return func(*args, **kwargs) 2025-12-04T09:41:20.8405390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 477, in forward 2025-12-04T09:41:20.8405794Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:20.8406183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8406549Z return func(*args, **kwargs) 2025-12-04T09:41:20.8406923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8407333Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8407773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8408228Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8408408Z 2025-12-04T09:41:20.8408484Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8408686Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8408878Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8409078Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8409276Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8409466Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8409664Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8409861Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8410047Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8410272Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8410613Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8410921Z res = mod(**inputs) 2025-12-04T09:41:20.8411265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8411634Z outputs = self.model( 2025-12-04T09:41:20.8411988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8412364Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8412746Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8413160Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8413501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8413844Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8414212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8414584Z return func(*args, **kwargs) 2025-12-04T09:41:20.8414942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 494, in forward 2025-12-04T09:41:20.8415363Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-12-04T09:41:20.8415772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8416139Z return func(*args, **kwargs) 2025-12-04T09:41:20.8416509Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8416916Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8417353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8417822Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8418002Z 2025-12-04T09:41:20.8418084Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8418296Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8418505Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8418704Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8418910Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8419115Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8419312Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8419521Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8419729Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8419937Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8420138Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8420344Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8420548Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8420773Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8421133Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8421452Z res = mod(**inputs) 2025-12-04T09:41:20.8421810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8422193Z outputs = self.model( 2025-12-04T09:41:20.8422561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8422946Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8423321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8423705Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8424052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8424410Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8424782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8425148Z return func(*args, **kwargs) 2025-12-04T09:41:20.8425519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 477, in forward 2025-12-04T09:41:20.8425918Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:20.8426343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8426729Z return func(*args, **kwargs) 2025-12-04T09:41:20.8427106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8427505Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8427958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8428421Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8428597Z 2025-12-04T09:41:20.8428681Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8428880Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8429109Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8429460Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8429771Z res = mod(**inputs) 2025-12-04T09:41:20.8430278Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8430666Z outputs = self.model( 2025-12-04T09:41:20.8431020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8431408Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8431783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8432159Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8432497Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8432856Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8433236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8433602Z return func(*args, **kwargs) 2025-12-04T09:41:20.8433961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 486, in forward 2025-12-04T09:41:20.8434349Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8434483Z 2025-12-04T09:41:20.8434568Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8434769Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8434977Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8435182Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8435376Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8435577Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8435774Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8436008Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8436345Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8436650Z res = mod(**inputs) 2025-12-04T09:41:20.8436999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8437355Z outputs = self.model( 2025-12-04T09:41:20.8437708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8438086Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8438457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8438826Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8439169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8439528Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8439986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8440346Z return func(*args, **kwargs) 2025-12-04T09:41:20.8440706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 494, in forward 2025-12-04T09:41:20.8441109Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-12-04T09:41:20.8441561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8441921Z return func(*args, **kwargs) 2025-12-04T09:41:20.8442280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8442722Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8443204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8443732Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8443938Z 2025-12-04T09:41:20.8444025Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8444224Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8444431Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8444634Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8444837Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8445033Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8445245Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8445442Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8445631Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8445829Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8446028Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8446215Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8446413Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8446643Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8446985Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8447302Z res = mod(**inputs) 2025-12-04T09:41:20.8447664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8448046Z outputs = self.model( 2025-12-04T09:41:20.8448406Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8448790Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8449168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8449547Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8449886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8450242Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8450616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8450984Z return func(*args, **kwargs) 2025-12-04T09:41:20.8451345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 477, in forward 2025-12-04T09:41:20.8451746Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:20.8452145Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8452503Z return func(*args, **kwargs) 2025-12-04T09:41:20.8452877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8453318Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8453734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8454197Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8454379Z 2025-12-04T09:41:20.8454458Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8454686Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8454884Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8455084Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8455284Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8455478Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8455676Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8455879Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8456079Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8456303Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8456659Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8456975Z res = mod(**inputs) 2025-12-04T09:41:20.8457329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8457710Z outputs = self.model( 2025-12-04T09:41:20.8458070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8458452Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8458817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8459199Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8459542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8459896Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8460272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8460634Z return func(*args, **kwargs) 2025-12-04T09:41:20.8461002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 494, in forward 2025-12-04T09:41:20.8461408Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-12-04T09:41:20.8461809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8462175Z return func(*args, **kwargs) 2025-12-04T09:41:20.8462533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8462936Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8463369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8463834Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8464008Z 2025-12-04T09:41:20.8464085Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8464291Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8464527Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8464885Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8464951Z res = mod(**inputs) 2025-12-04T09:41:20.8465201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8465275Z outputs = self.model( 2025-12-04T09:41:20.8465523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8465644Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8465907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8465977Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8466195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8466292Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8466521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8466595Z return func(*args, **kwargs) 2025-12-04T09:41:20.8466835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 503, in forward 2025-12-04T09:41:20.8466912Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8466923Z 2025-12-04T09:41:20.8467000Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8467074Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8467154Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8467225Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8467297Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8467376Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8467450Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8467521Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8467602Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8467674Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8467748Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8467856Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8468044Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8468113Z res = mod(**inputs) 2025-12-04T09:41:20.8468358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8468424Z outputs = self.model( 2025-12-04T09:41:20.8468670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8468740Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8468985Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8469053Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8469262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8469346Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8469575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8469643Z return func(*args, **kwargs) 2025-12-04T09:41:20.8469891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 477, in forward 2025-12-04T09:41:20.8469988Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:20.8470226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8470293Z return func(*args, **kwargs) 2025-12-04T09:41:20.8470530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8470630Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8470902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8471027Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8471045Z 2025-12-04T09:41:20.8471157Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8471235Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8471314Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8471387Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8471458Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8471537Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8471624Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8471696Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8471774Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8471874Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8472068Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8472131Z res = mod(**inputs) 2025-12-04T09:41:20.8472372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8472447Z outputs = self.model( 2025-12-04T09:41:20.8472686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8472756Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8473002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8473072Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8473286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8473363Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8473592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8473666Z return func(*args, **kwargs) 2025-12-04T09:41:20.8473905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 494, in forward 2025-12-04T09:41:20.8474010Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-12-04T09:41:20.8474245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8474312Z return func(*args, **kwargs) 2025-12-04T09:41:20.8474556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8474648Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8474918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8475046Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8475050Z 2025-12-04T09:41:20.8475122Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8475209Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8475281Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8475352Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8475431Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8475503Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8475603Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8475798Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8475861Z res = mod(**inputs) 2025-12-04T09:41:20.8476112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8476176Z outputs = self.model( 2025-12-04T09:41:20.8476414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8476517Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8476782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8476852Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8477070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8477163Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8477398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8477464Z return func(*args, **kwargs) 2025-12-04T09:41:20.8477702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 512, in forward 2025-12-04T09:41:20.8477787Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8477791Z 2025-12-04T09:41:20.8477863Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8477938Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8478016Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8478088Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8478165Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8478236Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8478306Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8478414Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8478600Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8478662Z res = mod(**inputs) 2025-12-04T09:41:20.8478910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8478975Z outputs = self.model( 2025-12-04T09:41:20.8479222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8479294Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8479531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8479608Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8479823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8479902Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8480144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8480212Z return func(*args, **kwargs) 2025-12-04T09:41:20.8480465Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 477, in forward 2025-12-04T09:41:20.8480565Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:20.8480805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8480881Z return func(*args, **kwargs) 2025-12-04T09:41:20.8481124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8481225Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8481572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8481705Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8481709Z 2025-12-04T09:41:20.8481793Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8481870Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8481948Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8482032Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8482130Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8482239Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8482316Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8482393Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8482475Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8482579Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8482776Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8482870Z res = mod(**inputs) 2025-12-04T09:41:20.8483126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8483194Z outputs = self.model( 2025-12-04T09:41:20.8483457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8483529Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8483791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8483862Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8484082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8484170Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8484413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8484490Z return func(*args, **kwargs) 2025-12-04T09:41:20.8484740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 494, in forward 2025-12-04T09:41:20.8484850Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-12-04T09:41:20.8485104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8485174Z return func(*args, **kwargs) 2025-12-04T09:41:20.8485415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8485515Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8485791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8485922Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8485925Z 2025-12-04T09:41:20.8486001Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8486076Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8486157Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8486231Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8486303Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8486386Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8486460Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8486541Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8486613Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8486686Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8486766Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8486838Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8486912Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8487018Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8487210Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8487273Z res = mod(**inputs) 2025-12-04T09:41:20.8487527Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8487594Z outputs = self.model( 2025-12-04T09:41:20.8487891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8487965Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8488210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8488285Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8488515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8488592Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8488833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8488900Z return func(*args, **kwargs) 2025-12-04T09:41:20.8489148Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 477, in forward 2025-12-04T09:41:20.8489246Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:20.8489483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8489564Z return func(*args, **kwargs) 2025-12-04T09:41:20.8489803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8489905Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8490180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8490302Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8490306Z 2025-12-04T09:41:20.8490387Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8490462Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8490560Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8490764Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8490826Z res = mod(**inputs) 2025-12-04T09:41:20.8491079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8491144Z outputs = self.model( 2025-12-04T09:41:20.8491389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8491467Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8491712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8491788Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8492001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8492078Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8492319Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8492388Z return func(*args, **kwargs) 2025-12-04T09:41:20.8492631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 486, in forward 2025-12-04T09:41:20.8492720Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8492723Z 2025-12-04T09:41:20.8492798Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8492878Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8492953Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8493028Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8493109Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8493181Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8493255Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8493412Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8493606Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8493669Z res = mod(**inputs) 2025-12-04T09:41:20.8493918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8494007Z outputs = self.model( 2025-12-04T09:41:20.8494264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8494336Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8494582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8494658Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8494873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8494958Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8495199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8495265Z return func(*args, **kwargs) 2025-12-04T09:41:20.8495513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 494, in forward 2025-12-04T09:41:20.8495619Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-12-04T09:41:20.8495849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8495921Z return func(*args, **kwargs) 2025-12-04T09:41:20.8496158Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8496255Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8496528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8496649Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8496652Z 2025-12-04T09:41:20.8496732Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8496804Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8496884Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8496956Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8497028Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8497108Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8497178Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8497248Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8497326Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8497396Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8497469Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8497549Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8497622Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8497719Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8497917Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8497980Z res = mod(**inputs) 2025-12-04T09:41:20.8498228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8498292Z outputs = self.model( 2025-12-04T09:41:20.8498530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8498608Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8498845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8498965Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8499179Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8499255Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8499491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8499573Z return func(*args, **kwargs) 2025-12-04T09:41:20.8499811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 477, in forward 2025-12-04T09:41:20.8499913Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:20.8500142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8500216Z return func(*args, **kwargs) 2025-12-04T09:41:20.8500456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8500550Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8500828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8500947Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8500952Z 2025-12-04T09:41:20.8501031Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8501104Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8501177Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8501256Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8501376Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8501446Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8501517Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8501597Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8501671Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8501769Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8501964Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8502025Z res = mod(**inputs) 2025-12-04T09:41:20.8502270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8502344Z outputs = self.model( 2025-12-04T09:41:20.8502578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8502655Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8502891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8502959Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8503179Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8503255Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8503489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8503555Z return func(*args, **kwargs) 2025-12-04T09:41:20.8503802Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 494, in forward 2025-12-04T09:41:20.8503915Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-12-04T09:41:20.8504143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8504209Z return func(*args, **kwargs) 2025-12-04T09:41:20.8504457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8504598Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8504877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8504998Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8505001Z 2025-12-04T09:41:20.8505074Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8505169Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8505268Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8505455Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8505526Z res = mod(**inputs) 2025-12-04T09:41:20.8505768Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8505840Z outputs = self.model( 2025-12-04T09:41:20.8506083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8506155Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8506403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8506470Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8506687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8506763Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8506989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8507061Z return func(*args, **kwargs) 2025-12-04T09:41:20.8507298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 503, in forward 2025-12-04T09:41:20.8507376Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8507380Z 2025-12-04T09:41:20.8507460Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8507531Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8507611Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8507684Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8507755Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8507836Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8507909Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8507979Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8508056Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8508126Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8508197Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8508302Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8508488Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8508560Z res = mod(**inputs) 2025-12-04T09:41:20.8508801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8508864Z outputs = self.model( 2025-12-04T09:41:20.8509108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8509178Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8509414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8509489Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8509696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8509778Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8510046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8510114Z return func(*args, **kwargs) 2025-12-04T09:41:20.8510362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 477, in forward 2025-12-04T09:41:20.8510459Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:20.8510708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8510777Z return func(*args, **kwargs) 2025-12-04T09:41:20.8511014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8511113Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8511394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8511520Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8511531Z 2025-12-04T09:41:20.8511605Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8511678Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8511758Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8511831Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8511904Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8511984Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8512056Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8512127Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8512206Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8512304Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8512501Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8512565Z res = mod(**inputs) 2025-12-04T09:41:20.8512808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8512883Z outputs = self.model( 2025-12-04T09:41:20.8513118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8513188Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8513436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8513505Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8513722Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8513799Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8514029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8514106Z return func(*args, **kwargs) 2025-12-04T09:41:20.8514354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 494, in forward 2025-12-04T09:41:20.8514459Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-12-04T09:41:20.8514694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8514761Z return func(*args, **kwargs) 2025-12-04T09:41:20.8515006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 322, in forward 2025-12-04T09:41:20.8515098Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:20.8515365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:20.8515490Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:20.8515515Z 2025-12-04T09:41:20.8516204Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8516292Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8516364Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8516437Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8516516Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8516589Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8516703Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8516901Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8516964Z res = mod(**inputs) 2025-12-04T09:41:20.8517204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1401, in forward 2025-12-04T09:41:20.8517279Z outputs = self.model( 2025-12-04T09:41:20.8517518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1276, in forward 2025-12-04T09:41:20.8517598Z decoder_outputs = self.decoder( 2025-12-04T09:41:20.8517835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1130, in forward 2025-12-04T09:41:20.8517903Z layer_outputs = decoder_layer( 2025-12-04T09:41:20.8518118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:20.8518196Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:20.8518431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:20.8518497Z return func(*args, **kwargs) 2025-12-04T09:41:20.8518734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 512, in forward 2025-12-04T09:41:20.8518818Z hidden_states = residual + hidden_states 2025-12-04T09:41:20.8518823Z 2025-12-04T09:41:20.8518899Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8518972Z cudagraph partition due to non gpu ops 2025-12-04T09:41:20.8519079Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:20.8519265Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:20.8519335Z res = mod(**inputs) 2025-12-04T09:41:20.8519578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1426, in forward 2025-12-04T09:41:20.8519740Z masked_lm_loss = loss_fct(lm_logits.view(-1, self.config.vocab_size), labels.view(-1)) 2025-12-04T09:41:20.8519743Z 2025-12-04T09:41:32.4583146Z Compilation time (from dynamo_timed): 31.566332674 2025-12-04T09:41:32.4624991Z pass 2025-12-04T09:41:32.4630975Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:41:32.4635773Z TIMING: _recursive_pre_grad_passes:0.06913 _recursive_joint_graph_passes:0.7834 _recursive_post_grad_passes:0.10944 async_compile.wait:0.84142 code_gen:11.63312 inductor_compile:14.52917 backend_compile:26.06986 gc:0.00016 entire_frame_compile:31.56633 total_wall_time:31.56633 2025-12-04T09:41:32.4637008Z STATS: call_* op count: 908 | FakeTensorMode.__torch_dispatch__:49173 | FakeTensor.__torch_dispatch__:6924 | ProxyTorchDispatchMode.__torch_dispatch__:10105 2025-12-04T09:41:32.4637586Z Dynamo produced 1 graphs covering 908 ops with 0 graph breaks (0 unique) 2025-12-04T09:41:35.5740611Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:41:35.5742544Z import pynvml # type: ignore[import] 2025-12-04T09:41:38.7943063Z 2025-12-04T09:41:40.9140724Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:41:40.9142404Z loading model: 0it [00:02, ?it/s] 2025-12-04T09:41:40.9142728Z cpu eval MBartForCausalLM 2025-12-04T09:41:42.5115310Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:41:42.8878426Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:41:43.2108755Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:41:52.5259626Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5260180Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5262958Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5263303Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5263655Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5263955Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5265099Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5269480Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5274171Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5278635Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5278936Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5279190Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5279421Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5279659Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5279886Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5280111Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5280330Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5280556Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5280779Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5281041Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5284103Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5284540Z res = mod(**inputs) 2025-12-04T09:41:52.5285010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5285473Z outputs = self.model.decoder( 2025-12-04T09:41:52.5285912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5286380Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5286844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5287265Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5287702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5288141Z return func(*args, **kwargs) 2025-12-04T09:41:52.5288572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 420, in forward 2025-12-04T09:41:52.5289031Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:52.5289483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5289942Z return func(*args, **kwargs) 2025-12-04T09:41:52.5290352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 267, in forward 2025-12-04T09:41:52.5290819Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:52.5291306Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:52.5291831Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:52.5292044Z 2025-12-04T09:41:52.5292137Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5292657Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5292988Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5293215Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5293439Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5293662Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5293880Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5294105Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5294376Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5294664Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5294889Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5295111Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5295332Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5295587Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5295990Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5296347Z res = mod(**inputs) 2025-12-04T09:41:52.5296748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5297178Z outputs = self.model.decoder( 2025-12-04T09:41:52.5297598Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5298017Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5298399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5298797Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5299210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5299611Z return func(*args, **kwargs) 2025-12-04T09:41:52.5300021Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 420, in forward 2025-12-04T09:41:52.5300474Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:52.5300906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5301301Z return func(*args, **kwargs) 2025-12-04T09:41:52.5301709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 267, in forward 2025-12-04T09:41:52.5302158Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:52.5302643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:52.5303158Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:52.5303362Z 2025-12-04T09:41:52.5303452Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5303686Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5303907Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5304137Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5304360Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5304574Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5304845Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5305220Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5305554Z res = mod(**inputs) 2025-12-04T09:41:52.5305944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5306364Z outputs = self.model.decoder( 2025-12-04T09:41:52.5306756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5307148Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5307560Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5307998Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5308394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5308775Z return func(*args, **kwargs) 2025-12-04T09:41:52.5309162Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 455, in forward 2025-12-04T09:41:52.5309588Z hidden_states = residual + hidden_states 2025-12-04T09:41:52.5309728Z 2025-12-04T09:41:52.5309816Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5310023Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5310237Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5310449Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5310653Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5310865Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5311078Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5311317Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5311692Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5312023Z res = mod(**inputs) 2025-12-04T09:41:52.5312400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5312793Z outputs = self.model.decoder( 2025-12-04T09:41:52.5313185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5313583Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5313935Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5314310Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5314707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5315087Z return func(*args, **kwargs) 2025-12-04T09:41:52.5315464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 420, in forward 2025-12-04T09:41:52.5315891Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:52.5316306Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5316675Z return func(*args, **kwargs) 2025-12-04T09:41:52.5317090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 267, in forward 2025-12-04T09:41:52.5317505Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:52.5317963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:52.5318455Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:52.5318652Z 2025-12-04T09:41:52.5318735Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5318956Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5319173Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5319379Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5319592Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5319805Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5320008Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5320225Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5320446Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5320663Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5320889Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5321112Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5321325Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5322530Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5322940Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5323297Z res = mod(**inputs) 2025-12-04T09:41:52.5323695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5324147Z outputs = self.model.decoder( 2025-12-04T09:41:52.5324564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5324983Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5325359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5325756Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5326169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5326574Z return func(*args, **kwargs) 2025-12-04T09:41:52.5326990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 420, in forward 2025-12-04T09:41:52.5327444Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:52.5327884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5328281Z return func(*args, **kwargs) 2025-12-04T09:41:52.5328690Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 267, in forward 2025-12-04T09:41:52.5329140Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:52.5329614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:52.5330500Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:52.5330709Z 2025-12-04T09:41:52.5330794Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5331013Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5331222Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5331438Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5331654Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5331866Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5332107Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5332475Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5332803Z res = mod(**inputs) 2025-12-04T09:41:52.5333162Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5333555Z outputs = self.model.decoder( 2025-12-04T09:41:52.5333944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5334322Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5334669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5335031Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5335407Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5335771Z return func(*args, **kwargs) 2025-12-04T09:41:52.5336144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 455, in forward 2025-12-04T09:41:52.5336538Z hidden_states = residual + hidden_states 2025-12-04T09:41:52.5336676Z 2025-12-04T09:41:52.5336755Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5336963Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5337238Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5337517Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5337720Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5337932Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5338138Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5338368Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5338732Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5339080Z res = mod(**inputs) 2025-12-04T09:41:52.5339439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5339831Z outputs = self.model.decoder( 2025-12-04T09:41:52.5340216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5340600Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5340948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5341308Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5341691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5342061Z return func(*args, **kwargs) 2025-12-04T09:41:52.5342427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 420, in forward 2025-12-04T09:41:52.5342837Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:52.5343237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5343600Z return func(*args, **kwargs) 2025-12-04T09:41:52.5343974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 267, in forward 2025-12-04T09:41:52.5344391Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:52.5344845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:52.5345306Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:52.5345490Z 2025-12-04T09:41:52.5345569Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5345780Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5345979Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5346181Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5346382Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5346582Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5346774Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5346975Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5347174Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5347368Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5347574Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5347773Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5347967Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5348199Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5348555Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5348871Z res = mod(**inputs) 2025-12-04T09:41:52.5349219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5349599Z outputs = self.model.decoder( 2025-12-04T09:41:52.5349972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5350346Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5350725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5351096Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5351468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5351827Z return func(*args, **kwargs) 2025-12-04T09:41:52.5352196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 420, in forward 2025-12-04T09:41:52.5352618Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:52.5353006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5353369Z return func(*args, **kwargs) 2025-12-04T09:41:52.5353739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 267, in forward 2025-12-04T09:41:52.5354142Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:52.5354569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:52.5355035Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:52.5355211Z 2025-12-04T09:41:52.5355297Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5355503Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5355701Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5355902Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5356107Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5356298Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5356528Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5356878Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5357186Z res = mod(**inputs) 2025-12-04T09:41:52.5357548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5357929Z outputs = self.model.decoder( 2025-12-04T09:41:52.5358301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5358669Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5359016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5359372Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5359737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5360107Z return func(*args, **kwargs) 2025-12-04T09:41:52.5360484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 455, in forward 2025-12-04T09:41:52.5360887Z hidden_states = residual + hidden_states 2025-12-04T09:41:52.5361025Z 2025-12-04T09:41:52.5361102Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5361313Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5361616Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5361836Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5362062Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5362290Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5362516Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5362765Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5363160Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5363485Z res = mod(**inputs) 2025-12-04T09:41:52.5363847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5364261Z outputs = self.model.decoder( 2025-12-04T09:41:52.5364677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5365064Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5365406Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5365782Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5366170Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5366536Z return func(*args, **kwargs) 2025-12-04T09:41:52.5366918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 420, in forward 2025-12-04T09:41:52.5367333Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:52.5367737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5368110Z return func(*args, **kwargs) 2025-12-04T09:41:52.5368489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 267, in forward 2025-12-04T09:41:52.5368902Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:52.5369345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:52.5369827Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:52.5370014Z 2025-12-04T09:41:52.5370097Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5370313Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5370519Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5370727Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5370937Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5371139Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5371349Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5371560Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5371761Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5371969Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5372178Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5372386Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5372592Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5372832Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5373194Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5373515Z res = mod(**inputs) 2025-12-04T09:41:52.5373885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5374288Z outputs = self.model.decoder( 2025-12-04T09:41:52.5374670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5375047Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5375393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5375755Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5376125Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5376494Z return func(*args, **kwargs) 2025-12-04T09:41:52.5376865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 420, in forward 2025-12-04T09:41:52.5377270Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:52.5377658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5378071Z return func(*args, **kwargs) 2025-12-04T09:41:52.5378431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 267, in forward 2025-12-04T09:41:52.5378828Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:52.5379270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:52.5379750Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:52.5379925Z 2025-12-04T09:41:52.5380011Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5380211Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5380414Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5380615Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5380808Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5381008Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5381243Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5381594Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5381898Z res = mod(**inputs) 2025-12-04T09:41:52.5382254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5382636Z outputs = self.model.decoder( 2025-12-04T09:41:52.5383002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5383374Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5383713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5384064Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5384427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5384791Z return func(*args, **kwargs) 2025-12-04T09:41:52.5385157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 455, in forward 2025-12-04T09:41:52.5385540Z hidden_states = residual + hidden_states 2025-12-04T09:41:52.5385673Z 2025-12-04T09:41:52.5385750Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5385955Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5386162Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5386360Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5386566Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5386768Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5386965Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5387198Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5387553Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5387881Z res = mod(**inputs) 2025-12-04T09:41:52.5388238Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5388624Z outputs = self.model.decoder( 2025-12-04T09:41:52.5389025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5389395Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5389734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5390086Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5390465Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5390831Z return func(*args, **kwargs) 2025-12-04T09:41:52.5391252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 420, in forward 2025-12-04T09:41:52.5391708Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:52.5392093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5392455Z return func(*args, **kwargs) 2025-12-04T09:41:52.5392820Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 267, in forward 2025-12-04T09:41:52.5393242Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:52.5393676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:52.5394146Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:52.5394332Z 2025-12-04T09:41:52.5394409Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5394617Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5394819Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5395025Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5395228Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5395422Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5395622Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5395823Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5396018Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5396217Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5396420Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5396613Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5396819Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5397048Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5397405Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5397711Z res = mod(**inputs) 2025-12-04T09:41:52.5398075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5398457Z outputs = self.model.decoder( 2025-12-04T09:41:52.5398825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5399205Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5399546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5399901Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5400267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5400626Z return func(*args, **kwargs) 2025-12-04T09:41:52.5400992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 420, in forward 2025-12-04T09:41:52.5401500Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:52.5401979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5402395Z return func(*args, **kwargs) 2025-12-04T09:41:52.5402796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 267, in forward 2025-12-04T09:41:52.5403202Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:52.5403701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:52.5404237Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:52.5404438Z 2025-12-04T09:41:52.5404540Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5404771Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5405030Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5405298Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5405522Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5405750Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5406015Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5406408Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5406787Z res = mod(**inputs) 2025-12-04T09:41:52.5407204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5407616Z outputs = self.model.decoder( 2025-12-04T09:41:52.5407982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5408366Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5408716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5409075Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5409442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5409814Z return func(*args, **kwargs) 2025-12-04T09:41:52.5410178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 455, in forward 2025-12-04T09:41:52.5410561Z hidden_states = residual + hidden_states 2025-12-04T09:41:52.5410703Z 2025-12-04T09:41:52.5410781Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5410991Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5411198Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5411395Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5411602Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5411807Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5412054Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5412284Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5412636Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5412956Z res = mod(**inputs) 2025-12-04T09:41:52.5413307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5413692Z outputs = self.model.decoder( 2025-12-04T09:41:52.5414068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5414451Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5414790Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5415150Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5415533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5415894Z return func(*args, **kwargs) 2025-12-04T09:41:52.5416266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 420, in forward 2025-12-04T09:41:52.5416682Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:52.5417069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5417415Z return func(*args, **kwargs) 2025-12-04T09:41:52.5417772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 267, in forward 2025-12-04T09:41:52.5418181Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:52.5418616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:52.5419115Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:52.5419296Z 2025-12-04T09:41:52.5419374Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5419581Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5419776Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5419979Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5420196Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5420398Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5420597Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5420793Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5420989Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5421179Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5421375Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5421570Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5421759Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5421985Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5422329Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5422627Z res = mod(**inputs) 2025-12-04T09:41:52.5422974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5423346Z outputs = self.model.decoder( 2025-12-04T09:41:52.5423708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5424072Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5424405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5424750Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5425105Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5425460Z return func(*args, **kwargs) 2025-12-04T09:41:52.5425817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 420, in forward 2025-12-04T09:41:52.5426215Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:41:52.5426597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5426961Z return func(*args, **kwargs) 2025-12-04T09:41:52.5427326Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 267, in forward 2025-12-04T09:41:52.5427730Z attn_output, attn_weights = attention_interface( 2025-12-04T09:41:52.5428156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:41:52.5428624Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:41:52.5428797Z 2025-12-04T09:41:52.5428884Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5429084Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5429287Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5429487Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5429685Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5429877Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5430239Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5430613Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5430924Z res = mod(**inputs) 2025-12-04T09:41:52.5431284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1865, in forward 2025-12-04T09:41:52.5431668Z outputs = self.model.decoder( 2025-12-04T09:41:52.5432134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1116, in forward 2025-12-04T09:41:52.5432535Z layer_outputs = decoder_layer( 2025-12-04T09:41:52.5432877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:41:52.5433234Z return super().__call__(*args, **kwargs) 2025-12-04T09:41:52.5433624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:41:52.5433987Z return func(*args, **kwargs) 2025-12-04T09:41:52.5434352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 455, in forward 2025-12-04T09:41:52.5434739Z hidden_states = residual + hidden_states 2025-12-04T09:41:52.5434872Z 2025-12-04T09:41:52.5434950Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5435154Z cudagraph partition due to non gpu ops 2025-12-04T09:41:52.5435389Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:41:52.5435730Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:41:52.5436046Z res = mod(**inputs) 2025-12-04T09:41:52.5436405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mbart/modeling_mbart.py", line 1887, in forward 2025-12-04T09:41:52.5436852Z loss = loss_fct(logits.view(-1, self.config.vocab_size), labels.view(-1)) 2025-12-04T09:41:52.5437043Z 2025-12-04T09:42:02.5306441Z Compilation time (from dynamo_timed): 18.35725445 2025-12-04T09:42:02.5562440Z pass 2025-12-04T09:42:02.5562976Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:42:02.5563979Z TIMING: _recursive_pre_grad_passes:0.02963 _recursive_joint_graph_passes:0.67662 _recursive_post_grad_passes:0.0604 async_compile.wait:1.03908 code_gen:9.4412 inductor_compile:11.34407 backend_compile:15.97965 gc:0.00117 entire_frame_compile:18.35725 total_wall_time:18.35725 2025-12-04T09:42:02.5566218Z STATS: call_* op count: 339 | FakeTensorMode.__torch_dispatch__:20500 | FakeTensor.__torch_dispatch__:2952 | ProxyTorchDispatchMode.__torch_dispatch__:4104 2025-12-04T09:42:02.5566761Z Dynamo produced 1 graphs covering 339 ops with 0 graph breaks (0 unique) 2025-12-04T09:42:04.9422243Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:42:04.9426978Z import pynvml # type: ignore[import] 2025-12-04T09:42:08.1567383Z 2025-12-04T09:42:10.2622890Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:42:10.2623476Z loading model: 0it [00:02, ?it/s] 2025-12-04T09:42:10.2623757Z cpu eval MT5ForConditionalGeneration 2025-12-04T09:42:10.8785991Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:42:11.1918122Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:42:11.5024390Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:42:26.7446676Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7450156Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7454579Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7456340Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7456577Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7456793Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7456994Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7465426Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7468697Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7469070Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7474063Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7478052Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7478478Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7478811Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7479110Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7479876Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7480704Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7481194Z res = mod(**inputs) 2025-12-04T09:42:26.7482198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1750, in forward 2025-12-04T09:42:26.7482780Z encoder_outputs = self.encoder( 2025-12-04T09:42:26.7483199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7483600Z layer_outputs = layer_module( 2025-12-04T09:42:26.7483988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7484378Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7484786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7485176Z return func(*args, **kwargs) 2025-12-04T09:42:26.7485554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7485966Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7486385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7486811Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7487225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7487622Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7487770Z 2025-12-04T09:42:26.7487867Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7488081Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7488291Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7488504Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7488711Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7488907Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7489111Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7489320Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7489523Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7489735Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7489952Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7490148Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7490352Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7490564Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7490765Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7490973Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7491220Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7491607Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7491941Z res = mod(**inputs) 2025-12-04T09:42:26.7492359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1750, in forward 2025-12-04T09:42:26.7492743Z encoder_outputs = self.encoder( 2025-12-04T09:42:26.7493115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7493492Z layer_outputs = layer_module( 2025-12-04T09:42:26.7493841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7494398Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7494778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7495166Z return func(*args, **kwargs) 2025-12-04T09:42:26.7495542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7495967Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7496369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7496816Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7497231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7497620Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7497773Z 2025-12-04T09:42:26.7497856Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7498069Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7498277Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7498477Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7498685Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7498893Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7499095Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7499301Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7499511Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7499708Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7499915Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7500117Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7500313Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7500517Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7500725Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7500932Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7501166Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7501529Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7501858Z res = mod(**inputs) 2025-12-04T09:42:26.7502210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1750, in forward 2025-12-04T09:42:26.7502591Z encoder_outputs = self.encoder( 2025-12-04T09:42:26.7502962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7503332Z layer_outputs = layer_module( 2025-12-04T09:42:26.7503674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7504036Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7504421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7504782Z return func(*args, **kwargs) 2025-12-04T09:42:26.7505147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7505532Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7505922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7506340Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7506739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7507119Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7507259Z 2025-12-04T09:42:26.7507339Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7507580Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7507820Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7508027Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7508227Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7508433Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7508638Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7508845Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7509069Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7509358Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7509567Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7509768Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7510004Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7510366Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7510688Z res = mod(**inputs) 2025-12-04T09:42:26.7511092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1750, in forward 2025-12-04T09:42:26.7511481Z encoder_outputs = self.encoder( 2025-12-04T09:42:26.7511857Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7512230Z layer_outputs = layer_module( 2025-12-04T09:42:26.7512579Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7512940Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7513317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7513694Z return func(*args, **kwargs) 2025-12-04T09:42:26.7514059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 565, in forward 2025-12-04T09:42:26.7514441Z self_attention_outputs = self.layer[0]( 2025-12-04T09:42:26.7514841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7515204Z return func(*args, **kwargs) 2025-12-04T09:42:26.7515568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 489, in forward 2025-12-04T09:42:26.7515994Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:42:26.7516189Z 2025-12-04T09:42:26.7516267Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7516476Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7516678Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7516886Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7517131Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7517524Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7517873Z res = mod(**inputs) 2025-12-04T09:42:26.7518259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1750, in forward 2025-12-04T09:42:26.7518669Z encoder_outputs = self.encoder( 2025-12-04T09:42:26.7519067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7519478Z layer_outputs = layer_module( 2025-12-04T09:42:26.7519855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7520246Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7520660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7521062Z return func(*args, **kwargs) 2025-12-04T09:42:26.7521518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7523023Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7523508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7523961Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7524419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7524869Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7525024Z 2025-12-04T09:42:26.7525110Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7525348Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7525572Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7525795Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7526011Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7526233Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7526459Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7526675Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7526899Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7527123Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7527337Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7527559Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7527785Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7527999Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7528223Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7528420Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7528648Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7528990Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7529304Z res = mod(**inputs) 2025-12-04T09:42:26.7529652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1750, in forward 2025-12-04T09:42:26.7530016Z encoder_outputs = self.encoder( 2025-12-04T09:42:26.7530588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7530957Z layer_outputs = layer_module( 2025-12-04T09:42:26.7531301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7531655Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7532034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7532405Z return func(*args, **kwargs) 2025-12-04T09:42:26.7532753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7533141Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7533527Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7533935Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7534334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7534715Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7534854Z 2025-12-04T09:42:26.7534938Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7535146Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7535346Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7535550Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7535753Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7535949Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7536150Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7536443Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7536686Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7536893Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7537098Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7537291Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7537529Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7537886Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7538230Z res = mod(**inputs) 2025-12-04T09:42:26.7538567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1750, in forward 2025-12-04T09:42:26.7538935Z encoder_outputs = self.encoder( 2025-12-04T09:42:26.7539297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7539652Z layer_outputs = layer_module( 2025-12-04T09:42:26.7539998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7540355Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7540728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7541074Z return func(*args, **kwargs) 2025-12-04T09:42:26.7541421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 565, in forward 2025-12-04T09:42:26.7541783Z self_attention_outputs = self.layer[0]( 2025-12-04T09:42:26.7542141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7542494Z return func(*args, **kwargs) 2025-12-04T09:42:26.7542835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 489, in forward 2025-12-04T09:42:26.7543245Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:42:26.7543422Z 2025-12-04T09:42:26.7543496Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7543696Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7543893Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7544081Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7544302Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7544645Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7544951Z res = mod(**inputs) 2025-12-04T09:42:26.7545278Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1750, in forward 2025-12-04T09:42:26.7545637Z encoder_outputs = self.encoder( 2025-12-04T09:42:26.7545988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7546336Z layer_outputs = layer_module( 2025-12-04T09:42:26.7546667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7547009Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7547370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7547713Z return func(*args, **kwargs) 2025-12-04T09:42:26.7548056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7548428Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7548799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7549184Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7549601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7549992Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7550128Z 2025-12-04T09:42:26.7550205Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7550407Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7550606Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7550821Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7551011Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7551209Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7551406Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7551595Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7551794Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7551991Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7552179Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7552374Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7552569Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7552756Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7552952Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7553150Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7553372Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7553703Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7554009Z res = mod(**inputs) 2025-12-04T09:42:26.7554341Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1750, in forward 2025-12-04T09:42:26.7554695Z encoder_outputs = self.encoder( 2025-12-04T09:42:26.7555046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7555400Z layer_outputs = layer_module( 2025-12-04T09:42:26.7555733Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7556075Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7556432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7556781Z return func(*args, **kwargs) 2025-12-04T09:42:26.7557120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7557491Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7557856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7558244Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7558624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7558987Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7559127Z 2025-12-04T09:42:26.7559210Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7559417Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7559613Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7559813Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7560013Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7560206Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7560406Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7560604Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7560796Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7560994Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7561222Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7561700Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7562063Z res = mod(**inputs) 2025-12-04T09:42:26.7562552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1787, in forward 2025-12-04T09:42:26.7563035Z decoder_outputs = self.decoder( 2025-12-04T09:42:26.7563444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7563857Z layer_outputs = layer_module( 2025-12-04T09:42:26.7564225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7564573Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7564951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7565314Z return func(*args, **kwargs) 2025-12-04T09:42:26.7565671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 565, in forward 2025-12-04T09:42:26.7566052Z self_attention_outputs = self.layer[0]( 2025-12-04T09:42:26.7566432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7566855Z return func(*args, **kwargs) 2025-12-04T09:42:26.7567207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 479, in forward 2025-12-04T09:42:26.7567579Z attention_output = self.SelfAttention( 2025-12-04T09:42:26.7567959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7568324Z return func(*args, **kwargs) 2025-12-04T09:42:26.7568669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 367, in forward 2025-12-04T09:42:26.7569041Z query_states = self.q(hidden_states) 2025-12-04T09:42:26.7569178Z 2025-12-04T09:42:26.7569259Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7569469Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7569669Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7569873Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7570076Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7570269Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7570473Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7570676Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7570868Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7571072Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7571273Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7571466Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7571667Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7571870Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7572098Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7572450Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7572764Z res = mod(**inputs) 2025-12-04T09:42:26.7573108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1750, in forward 2025-12-04T09:42:26.7573472Z encoder_outputs = self.encoder( 2025-12-04T09:42:26.7573834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7574201Z layer_outputs = layer_module( 2025-12-04T09:42:26.7574543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7574892Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7575263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7575626Z return func(*args, **kwargs) 2025-12-04T09:42:26.7576042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 565, in forward 2025-12-04T09:42:26.7576414Z self_attention_outputs = self.layer[0]( 2025-12-04T09:42:26.7576790Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7577145Z return func(*args, **kwargs) 2025-12-04T09:42:26.7577497Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 489, in forward 2025-12-04T09:42:26.7577910Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:42:26.7578087Z 2025-12-04T09:42:26.7578170Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7578371Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7578565Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7578765Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7578993Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7579332Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7579661Z res = mod(**inputs) 2025-12-04T09:42:26.7580003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1750, in forward 2025-12-04T09:42:26.7580354Z encoder_outputs = self.encoder( 2025-12-04T09:42:26.7580709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7581064Z layer_outputs = layer_module( 2025-12-04T09:42:26.7581396Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7581734Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7582094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7582451Z return func(*args, **kwargs) 2025-12-04T09:42:26.7582796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7583163Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7583534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7583929Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7584312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7584680Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7584822Z 2025-12-04T09:42:26.7584899Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7585103Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7585294Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7585494Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7585734Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7585930Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7586127Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7586324Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7586509Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7586704Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7586907Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7587095Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7587292Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7587487Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7587713Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7588056Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7588363Z res = mod(**inputs) 2025-12-04T09:42:26.7588756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1787, in forward 2025-12-04T09:42:26.7589111Z decoder_outputs = self.decoder( 2025-12-04T09:42:26.7589463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7589827Z layer_outputs = layer_module( 2025-12-04T09:42:26.7590160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7590515Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7590884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7591244Z return func(*args, **kwargs) 2025-12-04T09:42:26.7591590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7591969Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7592347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7592753Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7593146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7593526Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7593671Z 2025-12-04T09:42:26.7593749Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7593958Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7594156Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7594357Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7594559Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7594751Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7594953Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7595158Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7595352Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7595554Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7595754Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7595948Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7596153Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7596357Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7596557Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7596754Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7596957Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7597158Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7597349Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7597548Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7597752Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7597963Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7598224Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7614435Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7614709Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7614920Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7615160Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7615531Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7615870Z res = mod(**inputs) 2025-12-04T09:42:26.7616248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1787, in forward 2025-12-04T09:42:26.7616650Z decoder_outputs = self.decoder( 2025-12-04T09:42:26.7617034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7617406Z layer_outputs = layer_module( 2025-12-04T09:42:26.7617854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7618234Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7618619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7619001Z return func(*args, **kwargs) 2025-12-04T09:42:26.7619368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7619781Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7620183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7620597Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7620996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7621382Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7621535Z 2025-12-04T09:42:26.7621621Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7621837Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7622033Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7622239Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7622454Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7622643Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7622841Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7623038Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7623227Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7623424Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7623625Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7623823Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7624042Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7624395Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7624711Z res = mod(**inputs) 2025-12-04T09:42:26.7625051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1787, in forward 2025-12-04T09:42:26.7625427Z decoder_outputs = self.decoder( 2025-12-04T09:42:26.7625795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7626167Z layer_outputs = layer_module( 2025-12-04T09:42:26.7626507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7626868Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7627241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7627601Z return func(*args, **kwargs) 2025-12-04T09:42:26.7627965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 565, in forward 2025-12-04T09:42:26.7628338Z self_attention_outputs = self.layer[0]( 2025-12-04T09:42:26.7628716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7629073Z return func(*args, **kwargs) 2025-12-04T09:42:26.7629427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 489, in forward 2025-12-04T09:42:26.7629892Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:42:26.7630083Z 2025-12-04T09:42:26.7630302Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7630521Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7630732Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7630938Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7631219Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7631476Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7631689Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7631886Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7632096Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7632304Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7632504Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7632740Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7632946Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7633142Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7633381Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7633743Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7634069Z res = mod(**inputs) 2025-12-04T09:42:26.7634414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1787, in forward 2025-12-04T09:42:26.7634797Z decoder_outputs = self.decoder( 2025-12-04T09:42:26.7635169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7635531Z layer_outputs = layer_module( 2025-12-04T09:42:26.7635878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7636237Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7636616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7636971Z return func(*args, **kwargs) 2025-12-04T09:42:26.7637333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7637724Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7638112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7638515Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7638920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7639302Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7639449Z 2025-12-04T09:42:26.7639529Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7639742Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7639950Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7640153Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7640351Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7640552Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7640752Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7640948Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7641151Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7641356Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7641628Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7641835Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7642104Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7642299Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7642506Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7642702Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7642888Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7643081Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7643277Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7643477Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7643687Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7643897Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7644130Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7644588Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7644920Z res = mod(**inputs) 2025-12-04T09:42:26.7645281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1787, in forward 2025-12-04T09:42:26.7645641Z decoder_outputs = self.decoder( 2025-12-04T09:42:26.7646001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7646370Z layer_outputs = layer_module( 2025-12-04T09:42:26.7646692Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7647035Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7647396Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7647746Z return func(*args, **kwargs) 2025-12-04T09:42:26.7648086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 589, in forward 2025-12-04T09:42:26.7648448Z cross_attention_outputs = self.layer[1]( 2025-12-04T09:42:26.7648818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7649164Z return func(*args, **kwargs) 2025-12-04T09:42:26.7649498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 529, in forward 2025-12-04T09:42:26.7649904Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:42:26.7650074Z 2025-12-04T09:42:26.7650158Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7650352Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7650550Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7650748Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7650967Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7651311Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7651615Z res = mod(**inputs) 2025-12-04T09:42:26.7651946Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1787, in forward 2025-12-04T09:42:26.7652293Z decoder_outputs = self.decoder( 2025-12-04T09:42:26.7652646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7653002Z layer_outputs = layer_module( 2025-12-04T09:42:26.7653327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7653674Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7654037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7654393Z return func(*args, **kwargs) 2025-12-04T09:42:26.7654731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7655104Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7655470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7655863Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7656244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7656609Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7656743Z 2025-12-04T09:42:26.7656827Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7657019Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7657219Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7657462Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7657657Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7657849Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7658048Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7658243Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7658435Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7658648Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7658847Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7659033Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7659226Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7659419Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7659605Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7659800Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7659995Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7660185Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7660371Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7660567Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7660758Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7660943Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7661135Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7661328Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7661514Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7661707Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7661929Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7662263Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7662573Z res = mod(**inputs) 2025-12-04T09:42:26.7662909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1787, in forward 2025-12-04T09:42:26.7663266Z decoder_outputs = self.decoder( 2025-12-04T09:42:26.7663614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7663971Z layer_outputs = layer_module( 2025-12-04T09:42:26.7664303Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7664639Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7665002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7665353Z return func(*args, **kwargs) 2025-12-04T09:42:26.7665697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7666060Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7666428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7666822Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7667214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7667571Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7667711Z 2025-12-04T09:42:26.7667785Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7667990Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7668208Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7668549Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7668855Z res = mod(**inputs) 2025-12-04T09:42:26.7669192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1787, in forward 2025-12-04T09:42:26.7669552Z decoder_outputs = self.decoder( 2025-12-04T09:42:26.7669956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7670336Z layer_outputs = layer_module( 2025-12-04T09:42:26.7670667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7671020Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7671394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7671773Z return func(*args, **kwargs) 2025-12-04T09:42:26.7672131Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7672507Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7672882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 218, in forward 2025-12-04T09:42:26.7673303Z hidden_states = hidden_states + self.dropout(forwarded_states) 2025-12-04T09:42:26.7673484Z 2025-12-04T09:42:26.7673563Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7673775Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7673982Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7674177Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7674378Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7674587Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7674784Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7674989Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7675194Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7675390Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7675589Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7675795Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7675999Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7676195Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7676406Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7676608Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7676808Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7677014Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7677222Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7677419Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7677627Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7677835Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7678032Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7678241Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7678475Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7678831Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7679148Z res = mod(**inputs) 2025-12-04T09:42:26.7679502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1787, in forward 2025-12-04T09:42:26.7679877Z decoder_outputs = self.decoder( 2025-12-04T09:42:26.7680245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7680627Z layer_outputs = layer_module( 2025-12-04T09:42:26.7680980Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7681375Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7681888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7682308Z return func(*args, **kwargs) 2025-12-04T09:42:26.7682695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7683125Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7683568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7683981Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7684381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7684775Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7684916Z 2025-12-04T09:42:26.7684998Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7685197Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7685399Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7685600Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7685798Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7685988Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7686190Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7686393Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7686588Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7686791Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7686990Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7687186Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7687420Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7687774Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7688082Z res = mod(**inputs) 2025-12-04T09:42:26.7688428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1787, in forward 2025-12-04T09:42:26.7688799Z decoder_outputs = self.decoder( 2025-12-04T09:42:26.7689164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7689524Z layer_outputs = layer_module( 2025-12-04T09:42:26.7689871Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7690227Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7690601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7690959Z return func(*args, **kwargs) 2025-12-04T09:42:26.7691318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 565, in forward 2025-12-04T09:42:26.7691692Z self_attention_outputs = self.layer[0]( 2025-12-04T09:42:26.7692061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7692422Z return func(*args, **kwargs) 2025-12-04T09:42:26.7692776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 489, in forward 2025-12-04T09:42:26.7693195Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:42:26.7693373Z 2025-12-04T09:42:26.7693451Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7693661Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7693868Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7694063Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7694266Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7694470Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7694665Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7694867Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7695069Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7695267Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7695461Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7695660Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7695860Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7696073Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7696340Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7696692Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7697007Z res = mod(**inputs) 2025-12-04T09:42:26.7697337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1787, in forward 2025-12-04T09:42:26.7697727Z decoder_outputs = self.decoder( 2025-12-04T09:42:26.7698092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7698451Z layer_outputs = layer_module( 2025-12-04T09:42:26.7698784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7699127Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7699490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7699852Z return func(*args, **kwargs) 2025-12-04T09:42:26.7700193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7700561Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7700923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7701318Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7701708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7702077Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7702215Z 2025-12-04T09:42:26.7702292Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7702498Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7702704Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7702899Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7703099Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7703301Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7703494Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7703694Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7703894Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7704095Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7704298Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7704493Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7704686Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7704872Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7705067Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7705259Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7705447Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7705642Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7705838Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7706024Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7706218Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7706417Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7706643Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7706987Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7707307Z res = mod(**inputs) 2025-12-04T09:42:26.7707641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1787, in forward 2025-12-04T09:42:26.7707987Z decoder_outputs = self.decoder( 2025-12-04T09:42:26.7708340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7708693Z layer_outputs = layer_module( 2025-12-04T09:42:26.7709070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7709406Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7709763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7710110Z return func(*args, **kwargs) 2025-12-04T09:42:26.7710476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 589, in forward 2025-12-04T09:42:26.7710840Z cross_attention_outputs = self.layer[1]( 2025-12-04T09:42:26.7711201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7711555Z return func(*args, **kwargs) 2025-12-04T09:42:26.7711906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 529, in forward 2025-12-04T09:42:26.7712316Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:42:26.7712486Z 2025-12-04T09:42:26.7712566Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7712759Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7712952Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7713146Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7713361Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7713701Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7714010Z res = mod(**inputs) 2025-12-04T09:42:26.7714346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1787, in forward 2025-12-04T09:42:26.7714716Z decoder_outputs = self.decoder( 2025-12-04T09:42:26.7715075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1087, in forward 2025-12-04T09:42:26.7715450Z layer_outputs = layer_module( 2025-12-04T09:42:26.7715773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:42:26.7716117Z return super().__call__(*args, **kwargs) 2025-12-04T09:42:26.7716479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:42:26.7716831Z return func(*args, **kwargs) 2025-12-04T09:42:26.7717166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 615, in forward 2025-12-04T09:42:26.7717534Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:42:26.7717902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 217, in forward 2025-12-04T09:42:26.7718287Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-12-04T09:42:26.7718680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 186, in forward 2025-12-04T09:42:26.7719048Z hidden_states = hidden_gelu * hidden_linear 2025-12-04T09:42:26.7719180Z 2025-12-04T09:42:26.7719259Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7719452Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7719650Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7719846Z cudagraph partition due to non gpu ops 2025-12-04T09:42:26.7720059Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:42:26.7720399Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:42:26.7720702Z res = mod(**inputs) 2025-12-04T09:42:26.7721039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mt5/modeling_mt5.py", line 1823, in forward 2025-12-04T09:42:26.7721531Z loss = loss_fct(lm_logits.view(-1, lm_logits.size(-1)), labels.view(-1)) 2025-12-04T09:42:26.7721760Z 2025-12-04T09:42:36.8582474Z Compilation time (from dynamo_timed): 24.549141875 2025-12-04T09:42:36.8678542Z pass 2025-12-04T09:42:36.8678936Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:42:36.8679800Z TIMING: _recursive_pre_grad_passes:0.30385 _recursive_joint_graph_passes:1.12041 _recursive_post_grad_passes:0.08839 async_compile.wait:0.77765 code_gen:10.04685 inductor_compile:12.08328 backend_compile:20.54043 gc:0.00024 entire_frame_compile:24.54914 total_wall_time:24.54914 2025-12-04T09:42:36.8681278Z STATS: call_* op count: 1207 | FakeTensorMode.__torch_dispatch__:35742 | FakeTensor.__torch_dispatch__:5070 | ProxyTorchDispatchMode.__torch_dispatch__:9060 2025-12-04T09:42:36.8681927Z Dynamo produced 1 graphs covering 1207 ops with 0 graph breaks (0 unique) 2025-12-04T09:42:39.5289753Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:42:39.5290771Z import pynvml # type: ignore[import] 2025-12-04T09:42:42.7625720Z 2025-12-04T09:42:42.7745711Z loading model: 0it [00:00, ?it/s]If you want to use `MegatronBertForCausalLM` as a standalone, add `is_decoder=True.` 2025-12-04T09:42:42.7746441Z WARNING:transformers.models.megatron_bert.modeling_megatron_bert:If you want to use `MegatronBertForCausalLM` as a standalone, add `is_decoder=True.` 2025-12-04T09:42:45.2917110Z 2025-12-04T09:42:45.2918045Z loading model: 0it [00:02, ?it/s] 2025-12-04T09:42:45.2918482Z cpu eval MegatronBertForCausalLM 2025-12-04T09:42:46.8087776Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:42:47.1514822Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:42:47.4977690Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:43:05.6258840Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6259366Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6259830Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6260087Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6260485Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6260847Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6261399Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6261664Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6262040Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6262370Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6262622Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6262967Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6263187Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6263563Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6263774Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6269449Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6269714Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6269929Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6273845Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6274132Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6274375Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6274580Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6274788Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6274994Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6275195Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6275403Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6275620Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6275822Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6277040Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6277365Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6277576Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6277785Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6277992Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6278208Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6278404Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6278679Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6278884Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6279083Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6279328Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:05.6279724Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:05.6280066Z res = mod(**inputs) 2025-12-04T09:43:05.6280518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1056, in forward 2025-12-04T09:43:05.6280973Z outputs = self.bert( 2025-12-04T09:43:05.6281411Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 848, in forward 2025-12-04T09:43:05.6282087Z encoder_outputs = self.encoder( 2025-12-04T09:43:05.6282575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 532, in forward 2025-12-04T09:43:05.6283042Z layer_outputs = layer_module( 2025-12-04T09:43:05.6283411Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:43:05.6283797Z return super().__call__(*args, **kwargs) 2025-12-04T09:43:05.6284253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6284627Z return func(*args, **kwargs) 2025-12-04T09:43:05.6285053Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 439, in forward 2025-12-04T09:43:05.6285490Z self_attention_outputs = self.attention( 2025-12-04T09:43:05.6285882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6286259Z return func(*args, **kwargs) 2025-12-04T09:43:05.6286681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 375, in forward 2025-12-04T09:43:05.6287161Z attention_output = self.output(self_outputs[0], hidden_states) 2025-12-04T09:43:05.6287642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 324, in forward 2025-12-04T09:43:05.6288056Z return residual + hidden_states 2025-12-04T09:43:05.6288197Z 2025-12-04T09:43:05.6288283Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6288502Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6288707Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6288915Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6289127Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6289332Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6289543Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6289761Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6289966Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6290325Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6290537Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6290748Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6290969Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6291192Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6291429Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6291702Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6291958Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6292165Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6292362Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6292569Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6292777Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6292993Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6293213Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6293415Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6293616Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6293814Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6294019Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6294221Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6294455Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:05.6294823Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:05.6295155Z res = mod(**inputs) 2025-12-04T09:43:05.6295563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1056, in forward 2025-12-04T09:43:05.6295982Z outputs = self.bert( 2025-12-04T09:43:05.6296443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 848, in forward 2025-12-04T09:43:05.6296879Z encoder_outputs = self.encoder( 2025-12-04T09:43:05.6297290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 532, in forward 2025-12-04T09:43:05.6297715Z layer_outputs = layer_module( 2025-12-04T09:43:05.6298075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:43:05.6298447Z return super().__call__(*args, **kwargs) 2025-12-04T09:43:05.6298827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6299199Z return func(*args, **kwargs) 2025-12-04T09:43:05.6299630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 439, in forward 2025-12-04T09:43:05.6300065Z self_attention_outputs = self.attention( 2025-12-04T09:43:05.6300453Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6300837Z return func(*args, **kwargs) 2025-12-04T09:43:05.6301241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 375, in forward 2025-12-04T09:43:05.6301717Z attention_output = self.output(self_outputs[0], hidden_states) 2025-12-04T09:43:05.6302190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 324, in forward 2025-12-04T09:43:05.6302611Z return residual + hidden_states 2025-12-04T09:43:05.6302738Z 2025-12-04T09:43:05.6302819Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6303033Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6303243Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6303440Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6303647Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6303853Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6304051Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6304255Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6304457Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6304659Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6304856Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6305058Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6305288Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6305528Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6305735Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6305938Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6306137Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6306349Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6306559Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6306785Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6306997Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6307208Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6307418Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6307623Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6307835Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6308058Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6308258Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6308469Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6308720Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:05.6309091Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:05.6309429Z res = mod(**inputs) 2025-12-04T09:43:05.6309844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1056, in forward 2025-12-04T09:43:05.6310283Z outputs = self.bert( 2025-12-04T09:43:05.6310689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 848, in forward 2025-12-04T09:43:05.6311125Z encoder_outputs = self.encoder( 2025-12-04T09:43:05.6311555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 532, in forward 2025-12-04T09:43:05.6311989Z layer_outputs = layer_module( 2025-12-04T09:43:05.6312379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:43:05.6312766Z return super().__call__(*args, **kwargs) 2025-12-04T09:43:05.6313165Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6313574Z return func(*args, **kwargs) 2025-12-04T09:43:05.6314012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 439, in forward 2025-12-04T09:43:05.6314490Z self_attention_outputs = self.attention( 2025-12-04T09:43:05.6314897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6315281Z return func(*args, **kwargs) 2025-12-04T09:43:05.6315734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 375, in forward 2025-12-04T09:43:05.6316237Z attention_output = self.output(self_outputs[0], hidden_states) 2025-12-04T09:43:05.6316731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 324, in forward 2025-12-04T09:43:05.6317177Z return residual + hidden_states 2025-12-04T09:43:05.6317316Z 2025-12-04T09:43:05.6317400Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6317625Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6317840Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6318061Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6318280Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6318488Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6318705Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6318923Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6319137Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6319389Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6319677Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6319901Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6320114Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6320409Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6320621Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6320823Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6321060Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6321270Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6321577Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6321815Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6322042Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6322265Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6322480Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6322715Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6322930Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6323139Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6323364Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6323574Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6323804Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:05.6324175Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:05.6324508Z res = mod(**inputs) 2025-12-04T09:43:05.6324901Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1056, in forward 2025-12-04T09:43:05.6325323Z outputs = self.bert( 2025-12-04T09:43:05.6325723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 848, in forward 2025-12-04T09:43:05.6326144Z encoder_outputs = self.encoder( 2025-12-04T09:43:05.6326556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 532, in forward 2025-12-04T09:43:05.6326976Z layer_outputs = layer_module( 2025-12-04T09:43:05.6327329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:43:05.6327689Z return super().__call__(*args, **kwargs) 2025-12-04T09:43:05.6328060Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6328429Z return func(*args, **kwargs) 2025-12-04T09:43:05.6328836Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 439, in forward 2025-12-04T09:43:05.6329258Z self_attention_outputs = self.attention( 2025-12-04T09:43:05.6329641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6330009Z return func(*args, **kwargs) 2025-12-04T09:43:05.6330615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 375, in forward 2025-12-04T09:43:05.6331086Z attention_output = self.output(self_outputs[0], hidden_states) 2025-12-04T09:43:05.6331558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 324, in forward 2025-12-04T09:43:05.6331981Z return residual + hidden_states 2025-12-04T09:43:05.6332107Z 2025-12-04T09:43:05.6332197Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6332404Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6332614Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6332823Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6333024Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6333236Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6333503Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6333757Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6333968Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6334174Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6334381Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6334584Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6334788Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6335023Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6335218Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6335430Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6335626Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6335813Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6336006Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6336200Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6336387Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6336582Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6336780Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6336976Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6337163Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6337360Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6337556Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6337744Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6337969Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:05.6338314Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:05.6338614Z res = mod(**inputs) 2025-12-04T09:43:05.6338997Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1056, in forward 2025-12-04T09:43:05.6339392Z outputs = self.bert( 2025-12-04T09:43:05.6339774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 848, in forward 2025-12-04T09:43:05.6340168Z encoder_outputs = self.encoder( 2025-12-04T09:43:05.6340560Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 532, in forward 2025-12-04T09:43:05.6340960Z layer_outputs = layer_module( 2025-12-04T09:43:05.6341286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:43:05.6341631Z return super().__call__(*args, **kwargs) 2025-12-04T09:43:05.6341993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6342347Z return func(*args, **kwargs) 2025-12-04T09:43:05.6342730Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 439, in forward 2025-12-04T09:43:05.6343141Z self_attention_outputs = self.attention( 2025-12-04T09:43:05.6343507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6343855Z return func(*args, **kwargs) 2025-12-04T09:43:05.6344232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 375, in forward 2025-12-04T09:43:05.6344679Z attention_output = self.output(self_outputs[0], hidden_states) 2025-12-04T09:43:05.6345120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 324, in forward 2025-12-04T09:43:05.6345505Z return residual + hidden_states 2025-12-04T09:43:05.6345635Z 2025-12-04T09:43:05.6345710Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6345911Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6346110Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6346320Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6346571Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6346767Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6346955Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6347148Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6347343Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6347527Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6347746Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6347941Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6348127Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6348325Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6348522Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6348718Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6348908Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6349102Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6349297Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6349486Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6349686Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6349887Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6350076Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6350285Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6350480Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6350668Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6350863Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6351060Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6351288Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:05.6351630Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:05.6351946Z res = mod(**inputs) 2025-12-04T09:43:05.6352327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1056, in forward 2025-12-04T09:43:05.6352718Z outputs = self.bert( 2025-12-04T09:43:05.6353099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 848, in forward 2025-12-04T09:43:05.6353494Z encoder_outputs = self.encoder( 2025-12-04T09:43:05.6353891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 532, in forward 2025-12-04T09:43:05.6354281Z layer_outputs = layer_module( 2025-12-04T09:43:05.6354613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:43:05.6354956Z return super().__call__(*args, **kwargs) 2025-12-04T09:43:05.6355318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6355660Z return func(*args, **kwargs) 2025-12-04T09:43:05.6356058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 439, in forward 2025-12-04T09:43:05.6356478Z self_attention_outputs = self.attention( 2025-12-04T09:43:05.6356846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6357212Z return func(*args, **kwargs) 2025-12-04T09:43:05.6357610Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 375, in forward 2025-12-04T09:43:05.6358067Z attention_output = self.output(self_outputs[0], hidden_states) 2025-12-04T09:43:05.6358515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 324, in forward 2025-12-04T09:43:05.6358934Z return residual + hidden_states 2025-12-04T09:43:05.6359059Z 2025-12-04T09:43:05.6359170Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6359425Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6359636Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6359842Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6360047Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6360247Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6360458Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6360691Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6360900Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6361106Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6361310Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6361597Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6361820Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6362037Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6362255Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6362481Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6362713Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6362924Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6363122Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6363329Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6363536Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6363734Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6363947Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6364148Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6364341Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6364541Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6364744Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6364938Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6365172Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:05.6365534Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:05.6365859Z res = mod(**inputs) 2025-12-04T09:43:05.6366256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1056, in forward 2025-12-04T09:43:05.6366677Z outputs = self.bert( 2025-12-04T09:43:05.6367064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 848, in forward 2025-12-04T09:43:05.6367475Z encoder_outputs = self.encoder( 2025-12-04T09:43:05.6367871Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 532, in forward 2025-12-04T09:43:05.6368277Z layer_outputs = layer_module( 2025-12-04T09:43:05.6368617Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:43:05.6368961Z return super().__call__(*args, **kwargs) 2025-12-04T09:43:05.6369335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6369698Z return func(*args, **kwargs) 2025-12-04T09:43:05.6370097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 439, in forward 2025-12-04T09:43:05.6370507Z self_attention_outputs = self.attention( 2025-12-04T09:43:05.6370878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6371234Z return func(*args, **kwargs) 2025-12-04T09:43:05.6371626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 375, in forward 2025-12-04T09:43:05.6372083Z attention_output = self.output(self_outputs[0], hidden_states) 2025-12-04T09:43:05.6372600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 324, in forward 2025-12-04T09:43:05.6373029Z return residual + hidden_states 2025-12-04T09:43:05.6373150Z 2025-12-04T09:43:05.6373225Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6373435Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6373639Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6373833Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6374056Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6374260Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6374463Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6374656Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6374853Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6375052Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6375242Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6375439Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6375639Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6375834Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6376041Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6376238Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6376424Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6376620Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6376817Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6377017Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6377203Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6377399Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6377593Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6377784Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6377978Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6378173Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6378362Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6378559Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6378782Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:05.6379117Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:05.6379428Z res = mod(**inputs) 2025-12-04T09:43:05.6379812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1056, in forward 2025-12-04T09:43:05.6380212Z outputs = self.bert( 2025-12-04T09:43:05.6380587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 848, in forward 2025-12-04T09:43:05.6380987Z encoder_outputs = self.encoder( 2025-12-04T09:43:05.6381381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 532, in forward 2025-12-04T09:43:05.6381779Z layer_outputs = layer_module( 2025-12-04T09:43:05.6382106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:43:05.6382450Z return super().__call__(*args, **kwargs) 2025-12-04T09:43:05.6382810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6383155Z return func(*args, **kwargs) 2025-12-04T09:43:05.6383546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 439, in forward 2025-12-04T09:43:05.6383957Z self_attention_outputs = self.attention( 2025-12-04T09:43:05.6384320Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6384664Z return func(*args, **kwargs) 2025-12-04T09:43:05.6385055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 375, in forward 2025-12-04T09:43:05.6385560Z attention_output = self.output(self_outputs[0], hidden_states) 2025-12-04T09:43:05.6386016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 324, in forward 2025-12-04T09:43:05.6386413Z return residual + hidden_states 2025-12-04T09:43:05.6386539Z 2025-12-04T09:43:05.6386615Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6386840Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6387036Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6387241Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6387442Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6387635Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6387835Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6388032Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6388231Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6388425Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6388637Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6388831Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6389021Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6389222Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6389423Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6389616Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6389817Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6390018Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6390211Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6390419Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6390629Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6390845Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6391039Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6391240Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6391440Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6391633Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6391831Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6392032Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6392260Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:05.6392607Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:05.6392916Z res = mod(**inputs) 2025-12-04T09:43:05.6393301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1056, in forward 2025-12-04T09:43:05.6393701Z outputs = self.bert( 2025-12-04T09:43:05.6394090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 848, in forward 2025-12-04T09:43:05.6394500Z encoder_outputs = self.encoder( 2025-12-04T09:43:05.6394899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 532, in forward 2025-12-04T09:43:05.6395310Z layer_outputs = layer_module( 2025-12-04T09:43:05.6395648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:43:05.6395998Z return super().__call__(*args, **kwargs) 2025-12-04T09:43:05.6396361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6396722Z return func(*args, **kwargs) 2025-12-04T09:43:05.6397117Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 439, in forward 2025-12-04T09:43:05.6397532Z self_attention_outputs = self.attention( 2025-12-04T09:43:05.6397895Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6398278Z return func(*args, **kwargs) 2025-12-04T09:43:05.6398717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 375, in forward 2025-12-04T09:43:05.6399168Z attention_output = self.output(self_outputs[0], hidden_states) 2025-12-04T09:43:05.6399622Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 324, in forward 2025-12-04T09:43:05.6400051Z return residual + hidden_states 2025-12-04T09:43:05.6400172Z 2025-12-04T09:43:05.6400255Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6400452Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6400654Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6400857Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6401054Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6401256Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6401531Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6401740Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6401941Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6402141Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6402343Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6402539Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6402742Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6402955Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6403161Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6403384Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6403593Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6403792Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6403998Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6404202Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6404401Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6404615Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6404815Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6405020Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6405208Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6405406Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6405605Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6405799Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6406031Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:05.6406383Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:05.6406690Z res = mod(**inputs) 2025-12-04T09:43:05.6407081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1056, in forward 2025-12-04T09:43:05.6407499Z outputs = self.bert( 2025-12-04T09:43:05.6407877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 848, in forward 2025-12-04T09:43:05.6408270Z encoder_outputs = self.encoder( 2025-12-04T09:43:05.6408665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 532, in forward 2025-12-04T09:43:05.6409060Z layer_outputs = layer_module( 2025-12-04T09:43:05.6409388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:43:05.6409730Z return super().__call__(*args, **kwargs) 2025-12-04T09:43:05.6410096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6410458Z return func(*args, **kwargs) 2025-12-04T09:43:05.6410845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 439, in forward 2025-12-04T09:43:05.6411298Z self_attention_outputs = self.attention( 2025-12-04T09:43:05.6411710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6412075Z return func(*args, **kwargs) 2025-12-04T09:43:05.6412467Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 375, in forward 2025-12-04T09:43:05.6412956Z attention_output = self.output(self_outputs[0], hidden_states) 2025-12-04T09:43:05.6413404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 324, in forward 2025-12-04T09:43:05.6413805Z return residual + hidden_states 2025-12-04T09:43:05.6413934Z 2025-12-04T09:43:05.6414012Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6414217Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6414419Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6414615Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6414820Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6415022Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6415212Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6415412Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6415612Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6415813Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6416011Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6416206Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6416402Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6416590Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6416790Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6416991Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6417185Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6417383Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6417583Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6417775Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6417974Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6418175Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6418397Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6418590Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6418789Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6418991Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6419185Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6419384Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6419613Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:05.6419959Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:05.6420276Z res = mod(**inputs) 2025-12-04T09:43:05.6420668Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1056, in forward 2025-12-04T09:43:05.6421086Z outputs = self.bert( 2025-12-04T09:43:05.6421479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 848, in forward 2025-12-04T09:43:05.6421910Z encoder_outputs = self.encoder( 2025-12-04T09:43:05.6422318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 532, in forward 2025-12-04T09:43:05.6422737Z layer_outputs = layer_module( 2025-12-04T09:43:05.6423074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:43:05.6423431Z return super().__call__(*args, **kwargs) 2025-12-04T09:43:05.6423807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6424206Z return func(*args, **kwargs) 2025-12-04T09:43:05.6424692Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 439, in forward 2025-12-04T09:43:05.6425119Z self_attention_outputs = self.attention( 2025-12-04T09:43:05.6425493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:43:05.6425850Z return func(*args, **kwargs) 2025-12-04T09:43:05.6426268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 375, in forward 2025-12-04T09:43:05.6426730Z attention_output = self.output(self_outputs[0], hidden_states) 2025-12-04T09:43:05.6427189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 324, in forward 2025-12-04T09:43:05.6427588Z return residual + hidden_states 2025-12-04T09:43:05.6427714Z 2025-12-04T09:43:05.6427792Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6428002Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6428196Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6428396Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6428596Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6428787Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6428986Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6429187Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6429384Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6429575Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6429774Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6429974Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6430292Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6430504Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6430706Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6430905Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6431112Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6431315Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6431512Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6431715Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6431919Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6432124Z cudagraph partition due to non gpu ops 2025-12-04T09:43:05.6432353Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:05.6432707Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:05.6433025Z res = mod(**inputs) 2025-12-04T09:43:05.6433412Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1078, in forward 2025-12-04T09:43:05.6433826Z lm_loss = self.loss_function( 2025-12-04T09:43:05.6434187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/loss/loss_utils.py", line 67, in ForCausalLMLoss 2025-12-04T09:43:05.6434657Z loss = fixed_cross_entropy(logits, shift_labels, num_items_in_batch, ignore_index, **kwargs) 2025-12-04T09:43:05.6435124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/loss/loss_utils.py", line 36, in fixed_cross_entropy 2025-12-04T09:43:05.6435615Z loss = nn.functional.cross_entropy(source, target, ignore_index=ignore_index, reduction=reduction) 2025-12-04T09:43:05.6435857Z 2025-12-04T09:43:16.7439397Z Compilation time (from dynamo_timed): 28.477741076 2025-12-04T09:43:16.7464091Z pass 2025-12-04T09:43:16.7464969Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:43:16.7465925Z TIMING: _recursive_pre_grad_passes:0.05741 _recursive_joint_graph_passes:0.79441 _recursive_post_grad_passes:0.09834 async_compile.wait:0.78601 code_gen:10.56956 inductor_compile:13.01701 backend_compile:23.0787 gc:0.00072 entire_frame_compile:28.47774 total_wall_time:28.47774 2025-12-04T09:43:16.7467278Z STATS: call_* op count: 725 | FakeTensorMode.__torch_dispatch__:42167 | FakeTensor.__torch_dispatch__:5506 | ProxyTorchDispatchMode.__torch_dispatch__:9078 2025-12-04T09:43:16.7467825Z Dynamo produced 1 graphs covering 725 ops with 0 graph breaks (0 unique) 2025-12-04T09:43:19.7113898Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:43:19.7115080Z import pynvml # type: ignore[import] 2025-12-04T09:43:22.9786975Z 2025-12-04T09:43:23.4764721Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:43:23.4765031Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:43:23.4765290Z cpu eval MobileBertForMaskedLM 2025-12-04T09:43:23.7475004Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:43:23.9610848Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:43:24.3909606Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:43:58.7657918Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7663782Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7665604Z res = mod(**inputs) 2025-12-04T09:43:58.7666311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7672219Z outputs = self.mobilebert( 2025-12-04T09:43:58.7674624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 791, in forward 2025-12-04T09:43:58.7675348Z embedding_output = self.embeddings( 2025-12-04T09:43:58.7679291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 199, in forward 2025-12-04T09:43:58.7679849Z inputs_embeds = torch.cat( 2025-12-04T09:43:58.7680000Z 2025-12-04T09:43:58.7680140Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7680574Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7680949Z res = mod(**inputs) 2025-12-04T09:43:58.7681406Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7682054Z outputs = self.mobilebert( 2025-12-04T09:43:58.7682529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 791, in forward 2025-12-04T09:43:58.7683004Z embedding_output = self.embeddings( 2025-12-04T09:43:58.7683489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 208, in forward 2025-12-04T09:43:58.7684004Z inputs_embeds = self.embedding_transformation(inputs_embeds) 2025-12-04T09:43:58.7684220Z 2025-12-04T09:43:58.7684315Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7684561Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7684786Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7685020Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7685287Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7685688Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7686058Z res = mod(**inputs) 2025-12-04T09:43:58.7686496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7686978Z outputs = self.mobilebert( 2025-12-04T09:43:58.7688827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7689296Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7689757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7690216Z layer_outputs = layer_module( 2025-12-04T09:43:58.7690771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.7691263Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.7691738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.7692296Z self_outputs = self.self( 2025-12-04T09:43:58.7692751Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.7693216Z self.query(query_tensor) 2025-12-04T09:43:58.7693357Z 2025-12-04T09:43:58.7693449Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7693687Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7693912Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7694144Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7694387Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7694611Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7694823Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7695048Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7695272Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7695487Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7695710Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7695931Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7696145Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7696371Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7696591Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7696804Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7697060Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7697449Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7697800Z res = mod(**inputs) 2025-12-04T09:43:58.7698212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7698665Z outputs = self.mobilebert( 2025-12-04T09:43:58.7699099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7699552Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7699997Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7700444Z layer_outputs = layer_module( 2025-12-04T09:43:58.7700889Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.7701358Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.7701834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.7702332Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.7702821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7703268Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7703423Z 2025-12-04T09:43:58.7703503Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7703745Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7704039Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7704256Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7704466Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7704673Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7704878Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7705088Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7705350Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7705719Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7706054Z res = mod(**inputs) 2025-12-04T09:43:58.7706462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7706890Z outputs = self.mobilebert( 2025-12-04T09:43:58.7707295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7707737Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7708167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7708606Z layer_outputs = layer_module( 2025-12-04T09:43:58.7709046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.7709560Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.7710064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7710492Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7710646Z 2025-12-04T09:43:58.7710725Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7710941Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7711148Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7711363Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7711605Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7711969Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7712294Z res = mod(**inputs) 2025-12-04T09:43:58.7712699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7713127Z outputs = self.mobilebert( 2025-12-04T09:43:58.7713529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7713953Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7714370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7714793Z layer_outputs = layer_module( 2025-12-04T09:43:58.7715210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.7715730Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.7716245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.7716726Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.7717198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.7717642Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.7717787Z 2025-12-04T09:43:58.7717877Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7718086Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7718384Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7718758Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7719090Z res = mod(**inputs) 2025-12-04T09:43:58.7719482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7719929Z outputs = self.mobilebert( 2025-12-04T09:43:58.7720345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7720769Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7721192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7721718Z layer_outputs = layer_module( 2025-12-04T09:43:58.7722152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 496, in forward 2025-12-04T09:43:58.7722730Z query_tensor, key_tensor, value_tensor, layer_input = self.bottleneck(hidden_states) 2025-12-04T09:43:58.7723273Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 440, in forward 2025-12-04T09:43:58.7723753Z bottlenecked_hidden_states = self.input(hidden_states) 2025-12-04T09:43:58.7724219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 409, in forward 2025-12-04T09:43:58.7724638Z layer_input = self.dense(hidden_states) 2025-12-04T09:43:58.7724791Z 2025-12-04T09:43:58.7724873Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7725090Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7725322Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7725691Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7726018Z res = mod(**inputs) 2025-12-04T09:43:58.7726420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7726844Z outputs = self.mobilebert( 2025-12-04T09:43:58.7727269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7727724Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7728164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7728581Z layer_outputs = layer_module( 2025-12-04T09:43:58.7728998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.7729441Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.7729875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.7730496Z self_outputs = self.self( 2025-12-04T09:43:58.7730925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.7731351Z self.query(query_tensor) 2025-12-04T09:43:58.7731476Z 2025-12-04T09:43:58.7731557Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7731777Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7731995Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7732202Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7732420Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7732631Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7732842Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7733045Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7733357Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7733571Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7733777Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7733988Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7734200Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7734403Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7734642Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7734863Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7735116Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7735511Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7735888Z res = mod(**inputs) 2025-12-04T09:43:58.7736284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7736702Z outputs = self.mobilebert( 2025-12-04T09:43:58.7737118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7737544Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7737956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7738381Z layer_outputs = layer_module( 2025-12-04T09:43:58.7738831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.7739314Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.7739784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.7740248Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.7740712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7741147Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7741290Z 2025-12-04T09:43:58.7741370Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7741584Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7741798Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7742004Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7742214Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7742423Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7742631Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7742842Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7743092Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7743490Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7743815Z res = mod(**inputs) 2025-12-04T09:43:58.7744232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7744678Z outputs = self.mobilebert( 2025-12-04T09:43:58.7745086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7745504Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7745919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7746346Z layer_outputs = layer_module( 2025-12-04T09:43:58.7746757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.7747236Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.7747737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7748271Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7748433Z 2025-12-04T09:43:58.7748518Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7748735Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7751696Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7752045Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7752284Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7752660Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7753001Z res = mod(**inputs) 2025-12-04T09:43:58.7753397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7753803Z outputs = self.mobilebert( 2025-12-04T09:43:58.7754213Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7754631Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7755032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7755462Z layer_outputs = layer_module( 2025-12-04T09:43:58.7755864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.7756357Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.7756845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.7757311Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.7757787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.7758219Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.7758426Z 2025-12-04T09:43:58.7758509Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7758727Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7758980Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7759184Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7759434Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7759803Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7760136Z res = mod(**inputs) 2025-12-04T09:43:58.7760549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7761000Z outputs = self.mobilebert( 2025-12-04T09:43:58.7761452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7762047Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7762501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7762971Z layer_outputs = layer_module( 2025-12-04T09:43:58.7763387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.7763823Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.7764271Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.7764695Z self_outputs = self.self( 2025-12-04T09:43:58.7765108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.7765583Z self.query(query_tensor) 2025-12-04T09:43:58.7765752Z 2025-12-04T09:43:58.7765840Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7766067Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7766283Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7766508Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7766736Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7766963Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7767165Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7767377Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7767584Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7767784Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7767994Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7768205Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7768407Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7768618Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7768830Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7769034Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7769272Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7769645Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7769980Z res = mod(**inputs) 2025-12-04T09:43:58.7770374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7770795Z outputs = self.mobilebert( 2025-12-04T09:43:58.7771205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7771623Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7772076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7772522Z layer_outputs = layer_module( 2025-12-04T09:43:58.7772959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.7773429Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.7773905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.7774382Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.7774848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7775277Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7775436Z 2025-12-04T09:43:58.7775520Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7775745Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7775966Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7776191Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7776415Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7776635Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7776849Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7777070Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7777326Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7777720Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7778091Z res = mod(**inputs) 2025-12-04T09:43:58.7778535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7778985Z outputs = self.mobilebert( 2025-12-04T09:43:58.7779425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7779927Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7780371Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7780834Z layer_outputs = layer_module( 2025-12-04T09:43:58.7781293Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.7781813Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.7782335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7782781Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7782938Z 2025-12-04T09:43:58.7783022Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7783251Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7783466Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7783698Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7783952Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7784340Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7784678Z res = mod(**inputs) 2025-12-04T09:43:58.7785092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7785535Z outputs = self.mobilebert( 2025-12-04T09:43:58.7785958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7786401Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7786839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7787280Z layer_outputs = layer_module( 2025-12-04T09:43:58.7787709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.7788258Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.7788804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.7789308Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.7789809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.7790285Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.7790435Z 2025-12-04T09:43:58.7790529Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7790748Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7791004Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7791393Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7791740Z res = mod(**inputs) 2025-12-04T09:43:58.7792149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7792585Z outputs = self.mobilebert( 2025-12-04T09:43:58.7793021Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7793456Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7793894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7794332Z layer_outputs = layer_module( 2025-12-04T09:43:58.7794767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 496, in forward 2025-12-04T09:43:58.7795350Z query_tensor, key_tensor, value_tensor, layer_input = self.bottleneck(hidden_states) 2025-12-04T09:43:58.7795903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 440, in forward 2025-12-04T09:43:58.7796402Z bottlenecked_hidden_states = self.input(hidden_states) 2025-12-04T09:43:58.7796946Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 409, in forward 2025-12-04T09:43:58.7797402Z layer_input = self.dense(hidden_states) 2025-12-04T09:43:58.7797558Z 2025-12-04T09:43:58.7797643Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7797890Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7798137Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7798523Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7798873Z res = mod(**inputs) 2025-12-04T09:43:58.7799299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7799735Z outputs = self.mobilebert( 2025-12-04T09:43:58.7800177Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7800624Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7801063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7801608Z layer_outputs = layer_module( 2025-12-04T09:43:58.7802087Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.7802591Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.7803052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.7803513Z self_outputs = self.self( 2025-12-04T09:43:58.7803972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.7804444Z self.query(query_tensor) 2025-12-04T09:43:58.7804578Z 2025-12-04T09:43:58.7804666Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7804906Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7805146Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7805375Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7805613Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7805845Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7806066Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7806297Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7806528Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7806760Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7806981Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7807211Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7807441Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7807664Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7807904Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7808137Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7808391Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7808799Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7809163Z res = mod(**inputs) 2025-12-04T09:43:58.7809602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7810058Z outputs = self.mobilebert( 2025-12-04T09:43:58.7810562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7811026Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7811475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7811932Z layer_outputs = layer_module( 2025-12-04T09:43:58.7812422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.7812905Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.7813351Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.7813824Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.7814302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7814747Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7814890Z 2025-12-04T09:43:58.7814972Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7815202Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7815420Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7815623Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7815849Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7816072Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7816286Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7816509Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7816766Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7817169Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7817499Z res = mod(**inputs) 2025-12-04T09:43:58.7817909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7818343Z outputs = self.mobilebert( 2025-12-04T09:43:58.7818753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7819183Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7819612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7820042Z layer_outputs = layer_module( 2025-12-04T09:43:58.7820456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.7820941Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.7821435Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7821883Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7822028Z 2025-12-04T09:43:58.7822349Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7822564Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7822777Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7822979Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7823222Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7823594Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7823928Z res = mod(**inputs) 2025-12-04T09:43:58.7824317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7824731Z outputs = self.mobilebert( 2025-12-04T09:43:58.7825137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7825607Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7826032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7826458Z layer_outputs = layer_module( 2025-12-04T09:43:58.7826870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.7827389Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.7827897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.7828369Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.7828835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.7829261Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.7829418Z 2025-12-04T09:43:58.7829502Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7829730Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7829949Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7830352Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7830628Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7831024Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7831369Z res = mod(**inputs) 2025-12-04T09:43:58.7831787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7832221Z outputs = self.mobilebert( 2025-12-04T09:43:58.7832627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7833062Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7833503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7833948Z layer_outputs = layer_module( 2025-12-04T09:43:58.7834378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.7834844Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.7835300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.7835744Z self_outputs = self.self( 2025-12-04T09:43:58.7836171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.7836612Z self.query(query_tensor) 2025-12-04T09:43:58.7836739Z 2025-12-04T09:43:58.7836833Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7837050Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7837275Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7837499Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7837712Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7837935Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7838157Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7838377Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7838588Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7838810Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7839031Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7839242Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7839461Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7839682Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7839966Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7840240Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7840509Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7840894Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7841243Z res = mod(**inputs) 2025-12-04T09:43:58.7841738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7842229Z outputs = self.mobilebert( 2025-12-04T09:43:58.7842679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7843152Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7843607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7844090Z layer_outputs = layer_module( 2025-12-04T09:43:58.7844537Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.7845022Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.7845489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.7845950Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.7846415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7846855Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7846998Z 2025-12-04T09:43:58.7847085Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7847295Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7847507Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7847723Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7847933Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7848145Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7848357Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7848568Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7848804Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7849178Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7849513Z res = mod(**inputs) 2025-12-04T09:43:58.7849911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7850341Z outputs = self.mobilebert( 2025-12-04T09:43:58.7850761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7851189Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7851609Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7852034Z layer_outputs = layer_module( 2025-12-04T09:43:58.7852456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.7852919Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.7853390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7853828Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7853974Z 2025-12-04T09:43:58.7854061Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7854268Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7854485Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7854720Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7855610Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7855994Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7856330Z res = mod(**inputs) 2025-12-04T09:43:58.7856729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7857159Z outputs = self.mobilebert( 2025-12-04T09:43:58.7857572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7857996Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7858404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7858823Z layer_outputs = layer_module( 2025-12-04T09:43:58.7859243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.7859754Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.7860256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.7860729Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.7861202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.7861648Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.7861786Z 2025-12-04T09:43:58.7861864Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7862074Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7862308Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7862662Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7862983Z res = mod(**inputs) 2025-12-04T09:43:58.7863368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7863779Z outputs = self.mobilebert( 2025-12-04T09:43:58.7864170Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7864592Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7864998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7865406Z layer_outputs = layer_module( 2025-12-04T09:43:58.7865802Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 496, in forward 2025-12-04T09:43:58.7866299Z query_tensor, key_tensor, value_tensor, layer_input = self.bottleneck(hidden_states) 2025-12-04T09:43:58.7866804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 440, in forward 2025-12-04T09:43:58.7867329Z bottlenecked_hidden_states = self.input(hidden_states) 2025-12-04T09:43:58.7867776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 409, in forward 2025-12-04T09:43:58.7868194Z layer_input = self.dense(hidden_states) 2025-12-04T09:43:58.7868333Z 2025-12-04T09:43:58.7868420Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7868629Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7868873Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7869244Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7869580Z res = mod(**inputs) 2025-12-04T09:43:58.7870008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7870416Z outputs = self.mobilebert( 2025-12-04T09:43:58.7870820Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7871217Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7871639Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7872045Z layer_outputs = layer_module( 2025-12-04T09:43:58.7872451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.7872867Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.7873290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.7873700Z self_outputs = self.self( 2025-12-04T09:43:58.7874088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.7874493Z self.query(query_tensor) 2025-12-04T09:43:58.7874617Z 2025-12-04T09:43:58.7874695Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7874909Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7875111Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7875318Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7875526Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7875725Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7875941Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7876146Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7876347Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7876557Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7876769Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7876977Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7877182Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7877392Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7877599Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7877803Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7878047Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7878412Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7878736Z res = mod(**inputs) 2025-12-04T09:43:58.7879134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7879558Z outputs = self.mobilebert( 2025-12-04T09:43:58.7879971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7880397Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7880818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7881237Z layer_outputs = layer_module( 2025-12-04T09:43:58.7881737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.7882206Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.7882702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.7883216Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.7883720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7884245Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7884407Z 2025-12-04T09:43:58.7884492Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7884723Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7884949Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7885173Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7885413Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7885629Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7885850Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7886071Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7886317Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7886706Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7887056Z res = mod(**inputs) 2025-12-04T09:43:58.7887485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7887923Z outputs = self.mobilebert( 2025-12-04T09:43:58.7888371Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7888827Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7889267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7889710Z layer_outputs = layer_module( 2025-12-04T09:43:58.7890153Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.7890650Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.7891134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7891604Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7891754Z 2025-12-04T09:43:58.7891833Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7892047Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7892249Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7892459Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7892697Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7893056Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7893385Z res = mod(**inputs) 2025-12-04T09:43:58.7893780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7894196Z outputs = self.mobilebert( 2025-12-04T09:43:58.7894596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7895017Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7895429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7895837Z layer_outputs = layer_module( 2025-12-04T09:43:58.7896251Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.7896754Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.7897262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.7897729Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.7898205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.7898672Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.7898812Z 2025-12-04T09:43:58.7898898Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7899104Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7899316Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7899529Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7899766Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7900147Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7900471Z res = mod(**inputs) 2025-12-04T09:43:58.7900860Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7901260Z outputs = self.mobilebert( 2025-12-04T09:43:58.7901659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7902066Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7902463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7902880Z layer_outputs = layer_module( 2025-12-04T09:43:58.7903291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.7903726Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.7904149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.7904566Z self_outputs = self.self( 2025-12-04T09:43:58.7904972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.7905387Z self.query(query_tensor) 2025-12-04T09:43:58.7905509Z 2025-12-04T09:43:58.7905592Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7905806Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7906018Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7906226Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7906450Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7906654Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7906855Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7907059Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7907263Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7907460Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7907669Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7907880Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7908091Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7908294Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7908502Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7908712Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7908957Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7909346Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7909690Z res = mod(**inputs) 2025-12-04T09:43:58.7910075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7910493Z outputs = self.mobilebert( 2025-12-04T09:43:58.7910899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7911315Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7911718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7912134Z layer_outputs = layer_module( 2025-12-04T09:43:58.7912602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.7913050Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.7913489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.7913964Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.7914419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7914839Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7914987Z 2025-12-04T09:43:58.7915071Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7915283Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7915494Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7915703Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7915927Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7916150Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7916364Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7916583Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7916834Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7917207Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7917551Z res = mod(**inputs) 2025-12-04T09:43:58.7917965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7918403Z outputs = self.mobilebert( 2025-12-04T09:43:58.7918824Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7919265Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7919710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7920149Z layer_outputs = layer_module( 2025-12-04T09:43:58.7920576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.7921064Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.7921636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7922108Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7922267Z 2025-12-04T09:43:58.7922352Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7922580Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7922804Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7923018Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7923282Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7923647Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7923968Z res = mod(**inputs) 2025-12-04T09:43:58.7924365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7924780Z outputs = self.mobilebert( 2025-12-04T09:43:58.7925188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7925597Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7926010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7926425Z layer_outputs = layer_module( 2025-12-04T09:43:58.7926879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.7927407Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.7927915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.7928387Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.7928866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.7929301Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.7929448Z 2025-12-04T09:43:58.7929529Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7929742Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7929979Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7930527Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7930869Z res = mod(**inputs) 2025-12-04T09:43:58.7931262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7931687Z outputs = self.mobilebert( 2025-12-04T09:43:58.7932100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7932530Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7932940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7933366Z layer_outputs = layer_module( 2025-12-04T09:43:58.7933782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 496, in forward 2025-12-04T09:43:58.7934298Z query_tensor, key_tensor, value_tensor, layer_input = self.bottleneck(hidden_states) 2025-12-04T09:43:58.7934808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 440, in forward 2025-12-04T09:43:58.7935256Z bottlenecked_hidden_states = self.input(hidden_states) 2025-12-04T09:43:58.7935700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 409, in forward 2025-12-04T09:43:58.7936126Z layer_input = self.dense(hidden_states) 2025-12-04T09:43:58.7936270Z 2025-12-04T09:43:58.7936348Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7936560Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7936797Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7937144Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7937466Z res = mod(**inputs) 2025-12-04T09:43:58.7937868Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7938282Z outputs = self.mobilebert( 2025-12-04T09:43:58.7938706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7939115Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7939524Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7939928Z layer_outputs = layer_module( 2025-12-04T09:43:58.7940333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.7940760Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.7941180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.7941672Z self_outputs = self.self( 2025-12-04T09:43:58.7942080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.7942496Z self.query(query_tensor) 2025-12-04T09:43:58.7942615Z 2025-12-04T09:43:58.7942696Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7942933Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7943144Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7943354Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7943557Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7943765Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7943970Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7944171Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7944377Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7944584Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7944785Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7944991Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7945195Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7945393Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7945600Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7945809Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7946047Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7946399Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7946720Z res = mod(**inputs) 2025-12-04T09:43:58.7947107Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7947509Z outputs = self.mobilebert( 2025-12-04T09:43:58.7947913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7948334Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7948727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7949116Z layer_outputs = layer_module( 2025-12-04T09:43:58.7949516Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.7949937Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.7950348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.7950792Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.7951226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7951641Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7951776Z 2025-12-04T09:43:58.7951853Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7952055Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7952255Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7952457Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7952649Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7952849Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7953046Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7953241Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7953470Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7953821Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7954128Z res = mod(**inputs) 2025-12-04T09:43:58.7954506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7954959Z outputs = self.mobilebert( 2025-12-04T09:43:58.7955349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7955744Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7956137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7956598Z layer_outputs = layer_module( 2025-12-04T09:43:58.7956983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.7957427Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.7957873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7958284Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7958419Z 2025-12-04T09:43:58.7958497Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7958701Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7958903Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7959097Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7959327Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7959677Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7960023Z res = mod(**inputs) 2025-12-04T09:43:58.7960404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7960808Z outputs = self.mobilebert( 2025-12-04T09:43:58.7961210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7961742Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7962217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7962687Z layer_outputs = layer_module( 2025-12-04T09:43:58.7963130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.7963644Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.7964132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.7964581Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.7965029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.7965433Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.7965578Z 2025-12-04T09:43:58.7965656Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7965869Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7966071Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7966275Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7966505Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7966861Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7967167Z res = mod(**inputs) 2025-12-04T09:43:58.7967538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7967939Z outputs = self.mobilebert( 2025-12-04T09:43:58.7968318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7968760Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7969212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7969632Z layer_outputs = layer_module( 2025-12-04T09:43:58.7970027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.7970479Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.7970928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.7971352Z self_outputs = self.self( 2025-12-04T09:43:58.7971761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.7972190Z self.query(query_tensor) 2025-12-04T09:43:58.7972308Z 2025-12-04T09:43:58.7972399Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7972605Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7972814Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7973021Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7973228Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7973426Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7973633Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7973839Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7974035Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7974239Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7974442Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7974639Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7974845Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7975048Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7975244Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7975449Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7975688Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7976050Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7976373Z res = mod(**inputs) 2025-12-04T09:43:58.7976760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7977176Z outputs = self.mobilebert( 2025-12-04T09:43:58.7977569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7977984Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7978397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7978814Z layer_outputs = layer_module( 2025-12-04T09:43:58.7979217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.7979658Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.7980095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.7980545Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.7980991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7981416Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7981556Z 2025-12-04T09:43:58.7981641Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7981841Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7982048Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7982255Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7982491Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7982738Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7982946Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7983153Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7983378Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7983743Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7984080Z res = mod(**inputs) 2025-12-04T09:43:58.7984462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7984868Z outputs = self.mobilebert( 2025-12-04T09:43:58.7985270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7985674Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7986062Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7986458Z layer_outputs = layer_module( 2025-12-04T09:43:58.7986850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.7987286Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.7987725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.7988140Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.7988274Z 2025-12-04T09:43:58.7988359Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7988556Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7988758Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7988958Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7989177Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7989530Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7989842Z res = mod(**inputs) 2025-12-04T09:43:58.7990221Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7990625Z outputs = self.mobilebert( 2025-12-04T09:43:58.7991029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7991458Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7991869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7992286Z layer_outputs = layer_module( 2025-12-04T09:43:58.7992682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.7993166Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.7993644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.7994092Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.7994538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.7994948Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.7995083Z 2025-12-04T09:43:58.7995159Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7995365Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.7995597Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.7995937Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.7996275Z res = mod(**inputs) 2025-12-04T09:43:58.7996685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.7997088Z outputs = self.mobilebert( 2025-12-04T09:43:58.7997475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.7997887Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.7998282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.7998728Z layer_outputs = layer_module( 2025-12-04T09:43:58.7999114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 496, in forward 2025-12-04T09:43:58.7999599Z query_tensor, key_tensor, value_tensor, layer_input = self.bottleneck(hidden_states) 2025-12-04T09:43:58.8000092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 440, in forward 2025-12-04T09:43:58.8000525Z bottlenecked_hidden_states = self.input(hidden_states) 2025-12-04T09:43:58.8000954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 409, in forward 2025-12-04T09:43:58.8001361Z layer_input = self.dense(hidden_states) 2025-12-04T09:43:58.8001561Z 2025-12-04T09:43:58.8001655Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8001866Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8002110Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8002513Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8002882Z res = mod(**inputs) 2025-12-04T09:43:58.8003279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8003708Z outputs = self.mobilebert( 2025-12-04T09:43:58.8004127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8004516Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8004909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8005319Z layer_outputs = layer_module( 2025-12-04T09:43:58.8005742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8006167Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8006596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8007014Z self_outputs = self.self( 2025-12-04T09:43:58.8007413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8007826Z self.query(query_tensor) 2025-12-04T09:43:58.8007953Z 2025-12-04T09:43:58.8008033Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8008248Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8008455Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8008668Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8008880Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8009080Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8009287Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8009496Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8009698Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8009909Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8010148Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8010386Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8010592Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8010801Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8011013Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8011216Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8011459Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8011850Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8012189Z res = mod(**inputs) 2025-12-04T09:43:58.8012592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8013019Z outputs = self.mobilebert( 2025-12-04T09:43:58.8013435Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8013856Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8014281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8014682Z layer_outputs = layer_module( 2025-12-04T09:43:58.8015088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8015507Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8015934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8016374Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8016806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8017219Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8017366Z 2025-12-04T09:43:58.8017446Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8017650Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8017847Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8018051Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8018253Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8018444Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8018645Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8018849Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8019070Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8019421Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8019737Z res = mod(**inputs) 2025-12-04T09:43:58.8020114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8020520Z outputs = self.mobilebert( 2025-12-04T09:43:58.8020924Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8021338Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8021742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8022151Z layer_outputs = layer_module( 2025-12-04T09:43:58.8022556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8023015Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8023468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8023895Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8024063Z 2025-12-04T09:43:58.8024204Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8024412Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8024608Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8024812Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8025043Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8025402Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8025752Z res = mod(**inputs) 2025-12-04T09:43:58.8026139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8026550Z outputs = self.mobilebert( 2025-12-04T09:43:58.8026951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8027352Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8027752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8028141Z layer_outputs = layer_module( 2025-12-04T09:43:58.8028536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8029017Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8029497Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8029936Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8030553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8030978Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8031118Z 2025-12-04T09:43:58.8031209Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8031411Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8031619Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8031820Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8032046Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8032404Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8032723Z res = mod(**inputs) 2025-12-04T09:43:58.8033101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8033500Z outputs = self.mobilebert( 2025-12-04T09:43:58.8033894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8034298Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8034689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8035087Z layer_outputs = layer_module( 2025-12-04T09:43:58.8035481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8035894Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8036297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8036697Z self_outputs = self.self( 2025-12-04T09:43:58.8037087Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8037484Z self.query(query_tensor) 2025-12-04T09:43:58.8037598Z 2025-12-04T09:43:58.8037674Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8037919Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8038158Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8038354Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8038551Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8038750Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8038943Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8039144Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8039370Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8039564Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8039763Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8039960Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8040158Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8040347Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8040545Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8040741Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8040964Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8041327Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8041764Z res = mod(**inputs) 2025-12-04T09:43:58.8042208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8042687Z outputs = self.mobilebert( 2025-12-04T09:43:58.8043109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8043565Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8044017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8044495Z layer_outputs = layer_module( 2025-12-04T09:43:58.8044971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8045483Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8045973Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8046485Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8046986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8047458Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8047621Z 2025-12-04T09:43:58.8047708Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8047939Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8048170Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8048394Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8048620Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8048849Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8049076Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8049307Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8049546Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8049881Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8050186Z res = mod(**inputs) 2025-12-04T09:43:58.8050557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8050950Z outputs = self.mobilebert( 2025-12-04T09:43:58.8051322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8051713Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8052104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8052551Z layer_outputs = layer_module( 2025-12-04T09:43:58.8052928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8053351Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8053789Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8054209Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8054346Z 2025-12-04T09:43:58.8054418Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8054616Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8054819Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8055011Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8055238Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8055596Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8055891Z res = mod(**inputs) 2025-12-04T09:43:58.8056263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8056660Z outputs = self.mobilebert( 2025-12-04T09:43:58.8057043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8057430Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8057821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8058214Z layer_outputs = layer_module( 2025-12-04T09:43:58.8058596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8059079Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8059561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8060004Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8060444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8060853Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8060997Z 2025-12-04T09:43:58.8061072Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8061275Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8061496Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8061839Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8062152Z res = mod(**inputs) 2025-12-04T09:43:58.8062522Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8062915Z outputs = self.mobilebert( 2025-12-04T09:43:58.8063299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8063700Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8064088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8064484Z layer_outputs = layer_module( 2025-12-04T09:43:58.8064872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 496, in forward 2025-12-04T09:43:58.8065353Z query_tensor, key_tensor, value_tensor, layer_input = self.bottleneck(hidden_states) 2025-12-04T09:43:58.8065882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 440, in forward 2025-12-04T09:43:58.8066326Z bottlenecked_hidden_states = self.input(hidden_states) 2025-12-04T09:43:58.8066772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 409, in forward 2025-12-04T09:43:58.8067174Z layer_input = self.dense(hidden_states) 2025-12-04T09:43:58.8067333Z 2025-12-04T09:43:58.8067410Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8067617Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8067853Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8068211Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8068547Z res = mod(**inputs) 2025-12-04T09:43:58.8068929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8069339Z outputs = self.mobilebert( 2025-12-04T09:43:58.8069740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8070152Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8070555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8070963Z layer_outputs = layer_module( 2025-12-04T09:43:58.8071384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8071808Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8072238Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8072714Z self_outputs = self.self( 2025-12-04T09:43:58.8073119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8073526Z self.query(query_tensor) 2025-12-04T09:43:58.8073639Z 2025-12-04T09:43:58.8073716Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8073926Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8074131Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8074336Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8074535Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8074738Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8074938Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8075133Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8075334Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8075536Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8075731Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8075949Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8076151Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8076349Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8076554Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8076756Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8076984Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8077334Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8077659Z res = mod(**inputs) 2025-12-04T09:43:58.8078050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8078465Z outputs = self.mobilebert( 2025-12-04T09:43:58.8078876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8079302Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8079781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8080182Z layer_outputs = layer_module( 2025-12-04T09:43:58.8080587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8081035Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8081540Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8082002Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8082488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8082973Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8083135Z 2025-12-04T09:43:58.8083224Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8083463Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8083706Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8083915Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8084118Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8084330Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8084537Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8084736Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8084970Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8085332Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8085651Z res = mod(**inputs) 2025-12-04T09:43:58.8086047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8086463Z outputs = self.mobilebert( 2025-12-04T09:43:58.8086876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8087291Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8087709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8088132Z layer_outputs = layer_module( 2025-12-04T09:43:58.8088545Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8089018Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8089492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8089928Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8090071Z 2025-12-04T09:43:58.8090152Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8090369Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8090585Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8090790Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8091036Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8091405Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8091743Z res = mod(**inputs) 2025-12-04T09:43:58.8092126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8092539Z outputs = self.mobilebert( 2025-12-04T09:43:58.8092942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8093364Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8093798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8094224Z layer_outputs = layer_module( 2025-12-04T09:43:58.8094630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8095120Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8095637Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8096097Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8096555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8096976Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8097127Z 2025-12-04T09:43:58.8097209Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8097428Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8097645Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8097834Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8098057Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8098400Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8098708Z res = mod(**inputs) 2025-12-04T09:43:58.8099077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8099469Z outputs = self.mobilebert( 2025-12-04T09:43:58.8099845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8100237Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8100627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8101015Z layer_outputs = layer_module( 2025-12-04T09:43:58.8101392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8101801Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8102200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8102590Z self_outputs = self.self( 2025-12-04T09:43:58.8102961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8103345Z self.query(query_tensor) 2025-12-04T09:43:58.8103456Z 2025-12-04T09:43:58.8103538Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8103734Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8103934Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8104133Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8104328Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8104517Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8104596Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8104669Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8104741Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8104820Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8104891Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8104962Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8105043Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8105116Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8105188Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8105268Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8105370Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8105625Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8105689Z res = mod(**inputs) 2025-12-04T09:43:58.8105958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8106036Z outputs = self.mobilebert( 2025-12-04T09:43:58.8106311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8106382Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8106653Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8106723Z layer_outputs = layer_module( 2025-12-04T09:43:58.8106989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8107086Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8107347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8107464Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8107723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8107813Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8107817Z 2025-12-04T09:43:58.8107892Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8107964Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8108043Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8108115Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8108185Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8108268Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8108344Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8108416Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8108524Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8108710Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8108778Z res = mod(**inputs) 2025-12-04T09:43:58.8109038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8109105Z outputs = self.mobilebert( 2025-12-04T09:43:58.8109374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8109443Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8109713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8109785Z layer_outputs = layer_module( 2025-12-04T09:43:58.8110043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8110163Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8110422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8110504Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8110515Z 2025-12-04T09:43:58.8110589Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8110664Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8110741Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8110811Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8110910Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8111102Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8111215Z res = mod(**inputs) 2025-12-04T09:43:58.8111476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8111551Z outputs = self.mobilebert( 2025-12-04T09:43:58.8111810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8111905Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8112163Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8112229Z layer_outputs = layer_module( 2025-12-04T09:43:58.8112496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8112651Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8112922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8113040Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8113300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8113389Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8113394Z 2025-12-04T09:43:58.8113466Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8113537Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8113642Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8113825Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8113894Z res = mod(**inputs) 2025-12-04T09:43:58.8114166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8114234Z outputs = self.mobilebert( 2025-12-04T09:43:58.8114507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8114576Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8114854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8114921Z layer_outputs = layer_module( 2025-12-04T09:43:58.8115180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 496, in forward 2025-12-04T09:43:58.8115342Z query_tensor, key_tensor, value_tensor, layer_input = self.bottleneck(hidden_states) 2025-12-04T09:43:58.8115604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 440, in forward 2025-12-04T09:43:58.8115710Z bottlenecked_hidden_states = self.input(hidden_states) 2025-12-04T09:43:58.8115977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 409, in forward 2025-12-04T09:43:58.8116056Z layer_input = self.dense(hidden_states) 2025-12-04T09:43:58.8116061Z 2025-12-04T09:43:58.8116140Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8116212Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8116309Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8116501Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8116562Z res = mod(**inputs) 2025-12-04T09:43:58.8116826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8116914Z outputs = self.mobilebert( 2025-12-04T09:43:58.8117226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8117308Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8117575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8117659Z layer_outputs = layer_module( 2025-12-04T09:43:58.8117937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8118023Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8118308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8118380Z self_outputs = self.self( 2025-12-04T09:43:58.8118652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8118732Z self.query(query_tensor) 2025-12-04T09:43:58.8118735Z 2025-12-04T09:43:58.8118810Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8118886Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8118968Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8119044Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8119125Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8119199Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8119274Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8119357Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8119434Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8119513Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8119598Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8119672Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8119753Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8119841Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8119915Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8120000Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8120105Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8120301Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8120377Z res = mod(**inputs) 2025-12-04T09:43:58.8120654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8120726Z outputs = self.mobilebert( 2025-12-04T09:43:58.8121010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8121086Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8121380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8121453Z layer_outputs = layer_module( 2025-12-04T09:43:58.8121824Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8121952Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8122273Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8122413Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8122727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8122821Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8122825Z 2025-12-04T09:43:58.8122919Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8123068Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8123152Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8123249Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8123323Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8123395Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8123481Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8123576Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8123689Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8123881Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8123946Z res = mod(**inputs) 2025-12-04T09:43:58.8124223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8124293Z outputs = self.mobilebert( 2025-12-04T09:43:58.8124573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8124652Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8124930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8125007Z layer_outputs = layer_module( 2025-12-04T09:43:58.8125286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8125404Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8125695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8125780Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8125783Z 2025-12-04T09:43:58.8125865Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8125942Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8126022Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8126104Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8126219Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8126408Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8126481Z res = mod(**inputs) 2025-12-04T09:43:58.8126747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8126823Z outputs = self.mobilebert( 2025-12-04T09:43:58.8127091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8127161Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8127438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8127512Z layer_outputs = layer_module( 2025-12-04T09:43:58.8127786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8127951Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8128226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8128358Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8128633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8128717Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8128721Z 2025-12-04T09:43:58.8128804Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8128901Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8129013Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8129089Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8129190Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8129397Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8129460Z res = mod(**inputs) 2025-12-04T09:43:58.8129745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8129820Z outputs = self.mobilebert( 2025-12-04T09:43:58.8130086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8130316Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8130590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8130663Z layer_outputs = layer_module( 2025-12-04T09:43:58.8130941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8131027Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8131310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8131386Z self_outputs = self.self( 2025-12-04T09:43:58.8131661Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8131740Z self.query(query_tensor) 2025-12-04T09:43:58.8131744Z 2025-12-04T09:43:58.8131822Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8131897Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8131983Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8132065Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8132141Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8132226Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8132302Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8132389Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8132465Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8132541Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8132624Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8132700Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8132775Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8132859Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8132945Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8133022Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8133134Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8133336Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8133408Z res = mod(**inputs) 2025-12-04T09:43:58.8133684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8133753Z outputs = self.mobilebert( 2025-12-04T09:43:58.8134043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8134115Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8134383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8134458Z layer_outputs = layer_module( 2025-12-04T09:43:58.8134736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8134885Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8135200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8135312Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8135586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8135695Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8135698Z 2025-12-04T09:43:58.8135781Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8135856Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8135929Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8136010Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8136083Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8136156Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8136240Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8136316Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8136424Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8136619Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8136683Z res = mod(**inputs) 2025-12-04T09:43:58.8136955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8137025Z outputs = self.mobilebert( 2025-12-04T09:43:58.8137295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8137374Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8137640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8137716Z layer_outputs = layer_module( 2025-12-04T09:43:58.8137983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8138100Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8138371Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8138453Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8138456Z 2025-12-04T09:43:58.8138537Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8138609Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8138683Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8138762Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8138862Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8139049Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8139124Z res = mod(**inputs) 2025-12-04T09:43:58.8139388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8139456Z outputs = self.mobilebert( 2025-12-04T09:43:58.8139729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8139800Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8140076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8140144Z layer_outputs = layer_module( 2025-12-04T09:43:58.8140409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8140568Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8140906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8141035Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8141300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8141398Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8141403Z 2025-12-04T09:43:58.8141487Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8141562Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8141660Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8141855Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8141917Z res = mod(**inputs) 2025-12-04T09:43:58.8142195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8142265Z outputs = self.mobilebert( 2025-12-04T09:43:58.8142531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8142608Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8142876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8142953Z layer_outputs = layer_module( 2025-12-04T09:43:58.8143218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 496, in forward 2025-12-04T09:43:58.8143372Z query_tensor, key_tensor, value_tensor, layer_input = self.bottleneck(hidden_states) 2025-12-04T09:43:58.8143645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 440, in forward 2025-12-04T09:43:58.8143757Z bottlenecked_hidden_states = self.input(hidden_states) 2025-12-04T09:43:58.8144032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 409, in forward 2025-12-04T09:43:58.8144112Z layer_input = self.dense(hidden_states) 2025-12-04T09:43:58.8144115Z 2025-12-04T09:43:58.8144190Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8144272Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8144370Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8144563Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8144634Z res = mod(**inputs) 2025-12-04T09:43:58.8144901Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8144978Z outputs = self.mobilebert( 2025-12-04T09:43:58.8145249Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8145320Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8145592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8145660Z layer_outputs = layer_module( 2025-12-04T09:43:58.8145931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8146024Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8146294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8146370Z self_outputs = self.self( 2025-12-04T09:43:58.8146671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8146757Z self.query(query_tensor) 2025-12-04T09:43:58.8146761Z 2025-12-04T09:43:58.8146847Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8146923Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8147004Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8147079Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8147168Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8147248Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8147321Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8147393Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8147473Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8147545Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8147617Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8147697Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8147769Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8147846Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8147927Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8148000Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8148105Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8148293Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8148360Z res = mod(**inputs) 2025-12-04T09:43:58.8148632Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8148699Z outputs = self.mobilebert( 2025-12-04T09:43:58.8148964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8149042Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8149311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8149387Z layer_outputs = layer_module( 2025-12-04T09:43:58.8149653Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8149745Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8150019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8150130Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8172033Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8172264Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8172274Z 2025-12-04T09:43:58.8172386Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8172493Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8172586Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8172664Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8172740Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8172823Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8172898Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8172974Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8173105Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8173332Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8173402Z res = mod(**inputs) 2025-12-04T09:43:58.8173733Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8173814Z outputs = self.mobilebert( 2025-12-04T09:43:58.8174212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8174339Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8174629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8174716Z layer_outputs = layer_module( 2025-12-04T09:43:58.8174998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8175160Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8175438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8175527Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8175531Z 2025-12-04T09:43:58.8175621Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8175701Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8175785Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8175872Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8175984Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8176203Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8176270Z res = mod(**inputs) 2025-12-04T09:43:58.8176552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8176637Z outputs = self.mobilebert( 2025-12-04T09:43:58.8176918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8176996Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8177281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8177359Z layer_outputs = layer_module( 2025-12-04T09:43:58.8177643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8177811Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8178091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8178230Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8178514Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8178609Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8178614Z 2025-12-04T09:43:58.8178692Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8178769Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8178857Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8178932Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8179042Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8179256Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8179323Z res = mod(**inputs) 2025-12-04T09:43:58.8179618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8179696Z outputs = self.mobilebert( 2025-12-04T09:43:58.8179979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8180064Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8180346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8180502Z layer_outputs = layer_module( 2025-12-04T09:43:58.8180796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8180887Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8181178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8181272Z self_outputs = self.self( 2025-12-04T09:43:58.8181553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8181635Z self.query(query_tensor) 2025-12-04T09:43:58.8181639Z 2025-12-04T09:43:58.8181721Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8181806Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8181884Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8181964Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8182062Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8182138Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8182213Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8182296Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8182371Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8182446Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8182532Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8182607Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8182694Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8182771Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8182849Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8182936Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8183044Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8183246Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8183331Z res = mod(**inputs) 2025-12-04T09:43:58.8183642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8183720Z outputs = self.mobilebert( 2025-12-04T09:43:58.8184035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8184116Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8184431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8184510Z layer_outputs = layer_module( 2025-12-04T09:43:58.8184807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8184916Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8185205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8185334Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8185618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8185708Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8185711Z 2025-12-04T09:43:58.8185798Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8185876Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8185954Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8186040Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8186118Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8186203Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8186280Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8186375Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8186518Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8186723Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8186789Z res = mod(**inputs) 2025-12-04T09:43:58.8187078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8187167Z outputs = self.mobilebert( 2025-12-04T09:43:58.8187461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8187538Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8187821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8187903Z layer_outputs = layer_module( 2025-12-04T09:43:58.8188189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8188315Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8188604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8188692Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8188697Z 2025-12-04T09:43:58.8188784Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8188862Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8189092Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8189179Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8189288Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8189503Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8189573Z res = mod(**inputs) 2025-12-04T09:43:58.8189860Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8189942Z outputs = self.mobilebert( 2025-12-04T09:43:58.8190225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8190303Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8190620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8190698Z layer_outputs = layer_module( 2025-12-04T09:43:58.8191017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8191191Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8191503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8191655Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8191951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8192050Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8192055Z 2025-12-04T09:43:58.8192136Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8192215Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8192331Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8192532Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8192600Z res = mod(**inputs) 2025-12-04T09:43:58.8192892Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8193016Z outputs = self.mobilebert( 2025-12-04T09:43:58.8193309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8193384Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8193671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8193769Z layer_outputs = layer_module( 2025-12-04T09:43:58.8194050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 496, in forward 2025-12-04T09:43:58.8194226Z query_tensor, key_tensor, value_tensor, layer_input = self.bottleneck(hidden_states) 2025-12-04T09:43:58.8194514Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 440, in forward 2025-12-04T09:43:58.8194636Z bottlenecked_hidden_states = self.input(hidden_states) 2025-12-04T09:43:58.8194926Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 409, in forward 2025-12-04T09:43:58.8195010Z layer_input = self.dense(hidden_states) 2025-12-04T09:43:58.8195014Z 2025-12-04T09:43:58.8195093Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8195180Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8195284Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8195491Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8195557Z res = mod(**inputs) 2025-12-04T09:43:58.8195840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8195922Z outputs = self.mobilebert( 2025-12-04T09:43:58.8196209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8196286Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8196574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8196646Z layer_outputs = layer_module( 2025-12-04T09:43:58.8196937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8197027Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8197310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8197393Z self_outputs = self.self( 2025-12-04T09:43:58.8197676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8197757Z self.query(query_tensor) 2025-12-04T09:43:58.8197760Z 2025-12-04T09:43:58.8197837Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8197911Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8197992Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8198066Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8198140Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8198225Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8198300Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8198374Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8198457Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8198531Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8198609Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8198691Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8198766Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8198859Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8198984Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8199062Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8199164Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8199368Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8199431Z res = mod(**inputs) 2025-12-04T09:43:58.8199730Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8199802Z outputs = self.mobilebert( 2025-12-04T09:43:58.8200080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8200159Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8200436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8200517Z layer_outputs = layer_module( 2025-12-04T09:43:58.8200794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8200891Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8201174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8201291Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8201664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8201764Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8201768Z 2025-12-04T09:43:58.8201847Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8201933Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8202014Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8202093Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8202180Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8202257Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8202334Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8202420Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8202525Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8202736Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8202804Z res = mod(**inputs) 2025-12-04T09:43:58.8203092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8203171Z outputs = self.mobilebert( 2025-12-04T09:43:58.8203447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8203526Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8203816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8203888Z layer_outputs = layer_module( 2025-12-04T09:43:58.8204176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8204301Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8204582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8204677Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8204680Z 2025-12-04T09:43:58.8204758Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8204842Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8204960Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8205063Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8205173Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8205370Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8205434Z res = mod(**inputs) 2025-12-04T09:43:58.8205717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8205802Z outputs = self.mobilebert( 2025-12-04T09:43:58.8206085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8206166Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8206455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8206534Z layer_outputs = layer_module( 2025-12-04T09:43:58.8206812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8206972Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8207264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8207403Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8207678Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8207761Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8207764Z 2025-12-04T09:43:58.8207840Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8207923Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8207999Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8208076Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8208186Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8208387Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8208461Z res = mod(**inputs) 2025-12-04T09:43:58.8208731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8208804Z outputs = self.mobilebert( 2025-12-04T09:43:58.8209081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8209153Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8209430Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8209500Z layer_outputs = layer_module( 2025-12-04T09:43:58.8209785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8209879Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8210154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8210227Z self_outputs = self.self( 2025-12-04T09:43:58.8210511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8210582Z self.query(query_tensor) 2025-12-04T09:43:58.8210585Z 2025-12-04T09:43:58.8210670Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8210747Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8210822Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8210916Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8211008Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8211107Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8211191Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8211264Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8211337Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8211419Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8211492Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8211595Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8211668Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8211739Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8211818Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8211890Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8211990Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8212188Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8212253Z res = mod(**inputs) 2025-12-04T09:43:58.8212525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8212596Z outputs = self.mobilebert( 2025-12-04T09:43:58.8212865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8212948Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8213214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8213281Z layer_outputs = layer_module( 2025-12-04T09:43:58.8213554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8213648Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8213924Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8214034Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8214299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8214388Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8214393Z 2025-12-04T09:43:58.8214467Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8214549Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8214622Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8214694Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8214774Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8214846Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8214918Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8214997Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8215098Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8215286Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8215356Z res = mod(**inputs) 2025-12-04T09:43:58.8215625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8215703Z outputs = self.mobilebert( 2025-12-04T09:43:58.8215968Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8216037Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8216308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8216376Z layer_outputs = layer_module( 2025-12-04T09:43:58.8216668Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8216806Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8217073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8217161Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8217178Z 2025-12-04T09:43:58.8217252Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8217325Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8217406Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8217479Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8217585Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8217775Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8217837Z res = mod(**inputs) 2025-12-04T09:43:58.8218122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8218192Z outputs = self.mobilebert( 2025-12-04T09:43:58.8218466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8218545Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8218811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8218886Z layer_outputs = layer_module( 2025-12-04T09:43:58.8219151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8219306Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8219583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8219707Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8219981Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8220064Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8220069Z 2025-12-04T09:43:58.8220143Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8220226Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8220327Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8220516Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8220586Z res = mod(**inputs) 2025-12-04T09:43:58.8220857Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8220937Z outputs = self.mobilebert( 2025-12-04T09:43:58.8221214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8221286Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8221563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8221634Z layer_outputs = layer_module( 2025-12-04T09:43:58.8221901Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 496, in forward 2025-12-04T09:43:58.8222071Z query_tensor, key_tensor, value_tensor, layer_input = self.bottleneck(hidden_states) 2025-12-04T09:43:58.8222340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 440, in forward 2025-12-04T09:43:58.8222456Z bottlenecked_hidden_states = self.input(hidden_states) 2025-12-04T09:43:58.8222782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 409, in forward 2025-12-04T09:43:58.8222864Z layer_input = self.dense(hidden_states) 2025-12-04T09:43:58.8222868Z 2025-12-04T09:43:58.8222949Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8223022Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8223147Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8223336Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8223398Z res = mod(**inputs) 2025-12-04T09:43:58.8223670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8223741Z outputs = self.mobilebert( 2025-12-04T09:43:58.8224011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8224088Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8224354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8224430Z layer_outputs = layer_module( 2025-12-04T09:43:58.8224696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8224782Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8225055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8225125Z self_outputs = self.self( 2025-12-04T09:43:58.8225398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8225468Z self.query(query_tensor) 2025-12-04T09:43:58.8225472Z 2025-12-04T09:43:58.8225550Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8225633Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8225706Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8225779Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8225859Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8225931Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8226005Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8226085Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8226159Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8226238Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8226310Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8226383Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8226463Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8226535Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8226610Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8226692Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8226790Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8226978Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8227048Z res = mod(**inputs) 2025-12-04T09:43:58.8227315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8227393Z outputs = self.mobilebert( 2025-12-04T09:43:58.8227665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8227737Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8228018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8228106Z layer_outputs = layer_module( 2025-12-04T09:43:58.8228417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8228516Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8228788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8228924Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8229200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8229284Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8229296Z 2025-12-04T09:43:58.8229371Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8229447Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8229530Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8229610Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8229685Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8229766Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8229841Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8229915Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8230026Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8230479Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8230556Z res = mod(**inputs) 2025-12-04T09:43:58.8230837Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8230909Z outputs = self.mobilebert( 2025-12-04T09:43:58.8231199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8231273Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8231543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8231623Z layer_outputs = layer_module( 2025-12-04T09:43:58.8231899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8232029Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8232306Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8232391Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8232397Z 2025-12-04T09:43:58.8232483Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8232558Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8232641Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8232718Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8232826Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8233030Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8233096Z res = mod(**inputs) 2025-12-04T09:43:58.8233374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8233456Z outputs = self.mobilebert( 2025-12-04T09:43:58.8233728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8233808Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8234079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8234150Z layer_outputs = layer_module( 2025-12-04T09:43:58.8234533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8234693Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8234966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8235124Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8235406Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8235502Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8235506Z 2025-12-04T09:43:58.8235583Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8235662Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8235749Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8235827Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8235932Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8236139Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8236204Z res = mod(**inputs) 2025-12-04T09:43:58.8236491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8236564Z outputs = self.mobilebert( 2025-12-04T09:43:58.8236837Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8236913Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8237199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8237269Z layer_outputs = layer_module( 2025-12-04T09:43:58.8237552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8237639Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8237909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8237985Z self_outputs = self.self( 2025-12-04T09:43:58.8238257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8238327Z self.query(query_tensor) 2025-12-04T09:43:58.8238331Z 2025-12-04T09:43:58.8238413Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8238487Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8238568Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8238642Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8238717Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8238802Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8238879Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8238955Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8239036Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8239110Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8239183Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8239268Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8239341Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8239424Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8239497Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8239572Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8239680Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8239874Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8239939Z res = mod(**inputs) 2025-12-04T09:43:58.8240270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8240344Z outputs = self.mobilebert( 2025-12-04T09:43:58.8240619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8240699Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8240996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8241073Z layer_outputs = layer_module( 2025-12-04T09:43:58.8241347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8241442Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8241792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8241910Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8242191Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8242274Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8242280Z 2025-12-04T09:43:58.8242357Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8242442Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8242524Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8242604Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8242693Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8242774Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8242863Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8242946Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8243058Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8243278Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8243344Z res = mod(**inputs) 2025-12-04T09:43:58.8243616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8243699Z outputs = self.mobilebert( 2025-12-04T09:43:58.8243976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8244058Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8244336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8244408Z layer_outputs = layer_module( 2025-12-04T09:43:58.8244699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8244825Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8245101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8245193Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8245197Z 2025-12-04T09:43:58.8245275Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8245358Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8245435Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8245509Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8245620Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8245814Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8245879Z res = mod(**inputs) 2025-12-04T09:43:58.8246191Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8246279Z outputs = self.mobilebert( 2025-12-04T09:43:58.8246563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8246636Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8246911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8247008Z layer_outputs = layer_module( 2025-12-04T09:43:58.8247289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8247452Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8247726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8247855Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8248138Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8248222Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8248225Z 2025-12-04T09:43:58.8248311Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8248386Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8248488Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8248698Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8248759Z res = mod(**inputs) 2025-12-04T09:43:58.8249027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8249102Z outputs = self.mobilebert( 2025-12-04T09:43:58.8249375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8249452Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8249718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8249789Z layer_outputs = layer_module( 2025-12-04T09:43:58.8250062Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 496, in forward 2025-12-04T09:43:58.8250217Z query_tensor, key_tensor, value_tensor, layer_input = self.bottleneck(hidden_states) 2025-12-04T09:43:58.8250482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 440, in forward 2025-12-04T09:43:58.8250597Z bottlenecked_hidden_states = self.input(hidden_states) 2025-12-04T09:43:58.8250933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 409, in forward 2025-12-04T09:43:58.8251020Z layer_input = self.dense(hidden_states) 2025-12-04T09:43:58.8251024Z 2025-12-04T09:43:58.8251098Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8251170Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8251277Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8251465Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8251534Z res = mod(**inputs) 2025-12-04T09:43:58.8251799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8251866Z outputs = self.mobilebert( 2025-12-04T09:43:58.8252139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8252889Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8253184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8253262Z layer_outputs = layer_module( 2025-12-04T09:43:58.8253533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8253644Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8253912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8253981Z self_outputs = self.self( 2025-12-04T09:43:58.8254258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8254327Z self.query(query_tensor) 2025-12-04T09:43:58.8254333Z 2025-12-04T09:43:58.8254419Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8254494Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8254568Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8254650Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8254724Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8254798Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8254885Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8254959Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8255032Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8255114Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8255186Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8255267Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8255341Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8255415Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8255496Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8255573Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8255673Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8255870Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8255934Z res = mod(**inputs) 2025-12-04T09:43:58.8256199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8256279Z outputs = self.mobilebert( 2025-12-04T09:43:58.8256547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8256624Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8256891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8256960Z layer_outputs = layer_module( 2025-12-04T09:43:58.8257243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8257337Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8257611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8257721Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8257988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8258077Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8258080Z 2025-12-04T09:43:58.8258156Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8258229Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8258309Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8258400Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8258512Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8258586Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8258660Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8258738Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8258838Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8259026Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8259113Z res = mod(**inputs) 2025-12-04T09:43:58.8259381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8259458Z outputs = self.mobilebert( 2025-12-04T09:43:58.8259732Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8259804Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8260092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8260163Z layer_outputs = layer_module( 2025-12-04T09:43:58.8260437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8260563Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8260840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8260932Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8260935Z 2025-12-04T09:43:58.8261012Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8261087Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8261170Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8261247Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8261352Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8261565Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8261627Z res = mod(**inputs) 2025-12-04T09:43:58.8261902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8261973Z outputs = self.mobilebert( 2025-12-04T09:43:58.8262246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8262325Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8262596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8262671Z layer_outputs = layer_module( 2025-12-04T09:43:58.8262940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8263094Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8263373Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8263497Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8263765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8263854Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8263857Z 2025-12-04T09:43:58.8263932Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8264013Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8264086Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8264158Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8264316Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8264508Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8264570Z res = mod(**inputs) 2025-12-04T09:43:58.8264842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8264936Z outputs = self.mobilebert( 2025-12-04T09:43:58.8265209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8265288Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8265552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8265629Z layer_outputs = layer_module( 2025-12-04T09:43:58.8265897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8265985Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8266266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8266336Z self_outputs = self.self( 2025-12-04T09:43:58.8266611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8266687Z self.query(query_tensor) 2025-12-04T09:43:58.8266691Z 2025-12-04T09:43:58.8266768Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8266849Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8266925Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8266999Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8267080Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8267158Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8267235Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8267327Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8267401Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8267478Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8267550Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8267622Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8267704Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8267776Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8267850Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8267929Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8268028Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8268220Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8268290Z res = mod(**inputs) 2025-12-04T09:43:58.8268562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8268639Z outputs = self.mobilebert( 2025-12-04T09:43:58.8268909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8268979Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8269262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8269331Z layer_outputs = layer_module( 2025-12-04T09:43:58.8269606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8269710Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8269996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8270165Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8270448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8270534Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8270538Z 2025-12-04T09:43:58.8270620Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8270739Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8270821Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8270896Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8270972Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8271053Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8271131Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8271206Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8271315Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8271515Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8271579Z res = mod(**inputs) 2025-12-04T09:43:58.8271869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8271940Z outputs = self.mobilebert( 2025-12-04T09:43:58.8272231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8272308Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8272589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8272668Z layer_outputs = layer_module( 2025-12-04T09:43:58.8272961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8273090Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8273364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8273448Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8273451Z 2025-12-04T09:43:58.8273535Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8273611Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8273686Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8273767Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8273869Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8274071Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8274135Z res = mod(**inputs) 2025-12-04T09:43:58.8274409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8274489Z outputs = self.mobilebert( 2025-12-04T09:43:58.8274763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8274836Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8275115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8275187Z layer_outputs = layer_module( 2025-12-04T09:43:58.8275468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8275628Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8275920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8276099Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8276385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8276488Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8276491Z 2025-12-04T09:43:58.8276566Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8276660Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8276767Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8276958Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8277023Z res = mod(**inputs) 2025-12-04T09:43:58.8277314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8277387Z outputs = self.mobilebert( 2025-12-04T09:43:58.8277678Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8277752Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8278034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8278114Z layer_outputs = layer_module( 2025-12-04T09:43:58.8278395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 496, in forward 2025-12-04T09:43:58.8278566Z query_tensor, key_tensor, value_tensor, layer_input = self.bottleneck(hidden_states) 2025-12-04T09:43:58.8278909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 440, in forward 2025-12-04T09:43:58.8279029Z bottlenecked_hidden_states = self.input(hidden_states) 2025-12-04T09:43:58.8279341Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 409, in forward 2025-12-04T09:43:58.8279432Z layer_input = self.dense(hidden_states) 2025-12-04T09:43:58.8279436Z 2025-12-04T09:43:58.8279525Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8279607Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8279718Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8279933Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8280001Z res = mod(**inputs) 2025-12-04T09:43:58.8280307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8280391Z outputs = self.mobilebert( 2025-12-04T09:43:58.8280698Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8280785Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8281085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8281160Z layer_outputs = layer_module( 2025-12-04T09:43:58.8281463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 500, in forward 2025-12-04T09:43:58.8281706Z self_attention_outputs = self.attention( 2025-12-04T09:43:58.8282018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 334, in forward 2025-12-04T09:43:58.8282104Z self_outputs = self.self( 2025-12-04T09:43:58.8282424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 245, in forward 2025-12-04T09:43:58.8282509Z self.query(query_tensor) 2025-12-04T09:43:58.8282513Z 2025-12-04T09:43:58.8282637Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8282743Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8282831Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8282907Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8282983Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8283073Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8283150Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8283249Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8283327Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8283414Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8283497Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8283576Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8283651Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8283734Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8283808Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8283882Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8283992Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8284187Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8284259Z res = mod(**inputs) 2025-12-04T09:43:58.8284538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8284611Z outputs = self.mobilebert( 2025-12-04T09:43:58.8284891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8284964Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8285240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8285316Z layer_outputs = layer_module( 2025-12-04T09:43:58.8285591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 515, in forward 2025-12-04T09:43:58.8285693Z attention_output = ffn_module(attention_output) 2025-12-04T09:43:58.8285967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 469, in forward 2025-12-04T09:43:58.8286079Z intermediate_output = self.intermediate(hidden_states) 2025-12-04T09:43:58.8286365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8286447Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8286451Z 2025-12-04T09:43:58.8286537Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8286613Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8286690Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8286771Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8286847Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8286923Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8287004Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8287078Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8287181Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8287378Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8287444Z res = mod(**inputs) 2025-12-04T09:43:58.8287725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8287795Z outputs = self.mobilebert( 2025-12-04T09:43:58.8288072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8288152Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8288502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8288583Z layer_outputs = layer_module( 2025-12-04T09:43:58.8288860Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 518, in forward 2025-12-04T09:43:58.8288981Z intermediate_output = self.intermediate(attention_output) 2025-12-04T09:43:58.8289290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 359, in forward 2025-12-04T09:43:58.8289375Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8289379Z 2025-12-04T09:43:58.8289458Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8289545Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8289621Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8289708Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8289813Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8290016Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8290089Z res = mod(**inputs) 2025-12-04T09:43:58.8290376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 976, in forward 2025-12-04T09:43:58.8290449Z outputs = self.mobilebert( 2025-12-04T09:43:58.8290744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 794, in forward 2025-12-04T09:43:58.8290818Z encoder_outputs = self.encoder( 2025-12-04T09:43:58.8291111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 557, in forward 2025-12-04T09:43:58.8291188Z layer_outputs = layer_module( 2025-12-04T09:43:58.8291488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 519, in forward 2025-12-04T09:43:58.8291672Z layer_output = self.output(intermediate_output, attention_output, hidden_states) 2025-12-04T09:43:58.8291972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 398, in forward 2025-12-04T09:43:58.8292105Z layer_output = self.bottleneck(layer_output, residual_tensor_2) 2025-12-04T09:43:58.8292390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 372, in forward 2025-12-04T09:43:58.8292477Z layer_outputs = self.dense(hidden_states) 2025-12-04T09:43:58.8292480Z 2025-12-04T09:43:58.8292568Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8292647Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8292752Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8292962Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8293032Z res = mod(**inputs) 2025-12-04T09:43:58.8293327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 989, in forward 2025-12-04T09:43:58.8293428Z prediction_scores = self.cls(sequence_output) 2025-12-04T09:43:58.8293714Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 643, in forward 2025-12-04T09:43:58.8293840Z prediction_scores = self.predictions(sequence_output) 2025-12-04T09:43:58.8294127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 631, in forward 2025-12-04T09:43:58.8294230Z hidden_states = self.transform(hidden_states) 2025-12-04T09:43:58.8294519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 609, in forward 2025-12-04T09:43:58.8294623Z hidden_states = self.dense(hidden_states) 2025-12-04T09:43:58.8294657Z 2025-12-04T09:43:58.8294749Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8294827Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8294905Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8294989Z cudagraph partition due to non gpu ops 2025-12-04T09:43:58.8295093Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8295316Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8295381Z res = mod(**inputs) 2025-12-04T09:43:58.8295667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 994, in forward 2025-12-04T09:43:58.8295868Z masked_lm_loss = loss_fct(prediction_scores.view(-1, self.config.vocab_size), labels.view(-1)) 2025-12-04T09:43:58.8295872Z 2025-12-04T09:43:58.8295977Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:43:58.8296179Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:43:58.8296253Z res = mod(**inputs) 2025-12-04T09:43:58.8296537Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 989, in forward 2025-12-04T09:43:58.8296638Z prediction_scores = self.cls(sequence_output) 2025-12-04T09:43:58.8296923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 643, in forward 2025-12-04T09:43:58.8297037Z prediction_scores = self.predictions(sequence_output) 2025-12-04T09:43:58.8297333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 633, in forward 2025-12-04T09:43:58.8297414Z hidden_states += self.decoder.bias 2025-12-04T09:43:58.8297418Z 2025-12-04T09:44:11.4927052Z Compilation time (from dynamo_timed): 46.521724245 2025-12-04T09:44:11.4931506Z pass 2025-12-04T09:44:11.4934768Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:44:11.4937899Z TIMING: _recursive_pre_grad_passes:0.12307 _recursive_joint_graph_passes:1.39298 _recursive_post_grad_passes:0.202 async_compile.wait:0.8435 code_gen:11.55876 inductor_compile:15.86763 backend_compile:35.44996 gc:0.00027 entire_frame_compile:46.52172 total_wall_time:46.52172 2025-12-04T09:44:11.4938998Z STATS: call_* op count: 1451 | FakeTensorMode.__torch_dispatch__:82470 | FakeTensor.__torch_dispatch__:9495 | ProxyTorchDispatchMode.__torch_dispatch__:15219 2025-12-04T09:44:11.4942118Z Dynamo produced 1 graphs covering 1451 ops with 0 graph breaks (0 unique) 2025-12-04T09:44:15.2452148Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:44:15.2453092Z import pynvml # type: ignore[import] 2025-12-04T09:44:18.7445612Z 2025-12-04T09:44:20.0806867Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:44:20.0807266Z loading model: 0it [00:01, ?it/s] 2025-12-04T09:44:20.0807925Z cpu eval OPTForCausalLM 2025-12-04T09:44:22.0687083Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:44:22.4842576Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:44:22.8960205Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:44:32.3634491Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3634889Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3635145Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3635393Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3636036Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3636419Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3636703Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3636955Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3637195Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3637410Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3637612Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3638490Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3638695Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3638909Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3639120Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3639324Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3639538Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3639747Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3639949Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3640167Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3640455Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3640882Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3641285Z res = mod(**inputs) 2025-12-04T09:44:32.3641847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3642282Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3642720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3643138Z outputs = self.model.decoder( 2025-12-04T09:44:32.3643533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3643987Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3644374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3644755Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3645126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3645509Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3645903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3646296Z return func(*args, **kwargs) 2025-12-04T09:44:32.3646675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 255, in forward 2025-12-04T09:44:32.3647090Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:32.3647502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3647902Z return func(*args, **kwargs) 2025-12-04T09:44:32.3648284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 179, in forward 2025-12-04T09:44:32.3648696Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:32.3649187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:32.3649720Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:32.3649917Z 2025-12-04T09:44:32.3650010Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3650224Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3650471Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3650863Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3651219Z res = mod(**inputs) 2025-12-04T09:44:32.3651580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3652058Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3652475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3652879Z outputs = self.model.decoder( 2025-12-04T09:44:32.3653244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3653620Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3654003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3654382Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3654747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3655123Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3655511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3655895Z return func(*args, **kwargs) 2025-12-04T09:44:32.3656308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3656753Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3656931Z 2025-12-04T09:44:32.3657037Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3657412Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3657744Z res = mod(**inputs) 2025-12-04T09:44:32.3658082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3658437Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3658822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3659214Z outputs = self.model.decoder( 2025-12-04T09:44:32.3659584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3659987Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3660372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3660762Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3661117Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3661491Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3661885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3662258Z return func(*args, **kwargs) 2025-12-04T09:44:32.3662639Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3663048Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3663202Z 2025-12-04T09:44:32.3663317Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3663683Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3664016Z res = mod(**inputs) 2025-12-04T09:44:32.3664349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3664702Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3665084Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3665470Z outputs = self.model.decoder( 2025-12-04T09:44:32.3665821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3666230Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3666631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3667020Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3667384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3667768Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3668159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3668546Z return func(*args, **kwargs) 2025-12-04T09:44:32.3668911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3669308Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3669455Z 2025-12-04T09:44:32.3669566Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3669936Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3670256Z res = mod(**inputs) 2025-12-04T09:44:32.3670593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3670964Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3671339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3671723Z outputs = self.model.decoder( 2025-12-04T09:44:32.3672076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3672439Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3672810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3673195Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3673551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3673934Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3674312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3674690Z return func(*args, **kwargs) 2025-12-04T09:44:32.3675061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3675443Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3675593Z 2025-12-04T09:44:32.3675676Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3675897Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3676112Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3676314Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3676731Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3676942Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3677145Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3677357Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3677599Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3677965Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3678290Z res = mod(**inputs) 2025-12-04T09:44:32.3678620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3678982Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3679381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3679788Z outputs = self.model.decoder( 2025-12-04T09:44:32.3680228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3680621Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3681036Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3681450Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3681945Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3682368Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3682808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3683216Z return func(*args, **kwargs) 2025-12-04T09:44:32.3683603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 255, in forward 2025-12-04T09:44:32.3684043Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:32.3684484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3684893Z return func(*args, **kwargs) 2025-12-04T09:44:32.3685284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 179, in forward 2025-12-04T09:44:32.3685733Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:32.3686220Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:32.3686772Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:32.3686971Z 2025-12-04T09:44:32.3687060Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3687296Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3687555Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3687943Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3688295Z res = mod(**inputs) 2025-12-04T09:44:32.3688647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3689041Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3689442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3689850Z outputs = self.model.decoder( 2025-12-04T09:44:32.3690227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3690610Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3691014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3691425Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3691798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3692153Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3692533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3692911Z return func(*args, **kwargs) 2025-12-04T09:44:32.3693286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3693690Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3693857Z 2025-12-04T09:44:32.3693965Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3694335Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3694665Z res = mod(**inputs) 2025-12-04T09:44:32.3694988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3695399Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3695782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3696154Z outputs = self.model.decoder( 2025-12-04T09:44:32.3696507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3696884Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3697257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3697644Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3698004Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3698373Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3698758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3699140Z return func(*args, **kwargs) 2025-12-04T09:44:32.3699516Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3699917Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3700220Z 2025-12-04T09:44:32.3700332Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3700702Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3701030Z res = mod(**inputs) 2025-12-04T09:44:32.3701358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3701722Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3702104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3702493Z outputs = self.model.decoder( 2025-12-04T09:44:32.3702838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3703199Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3703582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3703961Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3704320Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3704704Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3705090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3705462Z return func(*args, **kwargs) 2025-12-04T09:44:32.3705836Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3706226Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3706365Z 2025-12-04T09:44:32.3706472Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3706837Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3707166Z res = mod(**inputs) 2025-12-04T09:44:32.3707496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3707849Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3708229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3708613Z outputs = self.model.decoder( 2025-12-04T09:44:32.3708966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3709378Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3709762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3710143Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3710495Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3710909Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3711411Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3711804Z return func(*args, **kwargs) 2025-12-04T09:44:32.3712176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3712571Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3712713Z 2025-12-04T09:44:32.3712807Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3713026Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3713245Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3713459Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3713670Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3713874Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3714085Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3714298Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3714531Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3714904Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3715235Z res = mod(**inputs) 2025-12-04T09:44:32.3715563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3715929Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3716320Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3716709Z outputs = self.model.decoder( 2025-12-04T09:44:32.3717058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3717419Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3717807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3718184Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3718546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3718918Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3719311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3719690Z return func(*args, **kwargs) 2025-12-04T09:44:32.3720099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 255, in forward 2025-12-04T09:44:32.3720545Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:32.3720991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3721436Z return func(*args, **kwargs) 2025-12-04T09:44:32.3721897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 179, in forward 2025-12-04T09:44:32.3722357Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:32.3722855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:32.3723400Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:32.3723641Z 2025-12-04T09:44:32.3723728Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3723996Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3724253Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3724648Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3725008Z res = mod(**inputs) 2025-12-04T09:44:32.3725356Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3725764Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3726169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3726587Z outputs = self.model.decoder( 2025-12-04T09:44:32.3726958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3727344Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3727753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3728150Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3728531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3728926Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3729338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3729745Z return func(*args, **kwargs) 2025-12-04T09:44:32.3730450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3730889Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3731044Z 2025-12-04T09:44:32.3731154Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3731518Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3731843Z res = mod(**inputs) 2025-12-04T09:44:32.3732168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3732519Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3732896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3733278Z outputs = self.model.decoder( 2025-12-04T09:44:32.3733626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3733976Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3734348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3734726Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3735074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3735438Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3735817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3736189Z return func(*args, **kwargs) 2025-12-04T09:44:32.3736547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3736941Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3737094Z 2025-12-04T09:44:32.3737207Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3737567Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3737882Z res = mod(**inputs) 2025-12-04T09:44:32.3738302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3738687Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3739055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3739433Z outputs = self.model.decoder( 2025-12-04T09:44:32.3739783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3740164Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3740529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3740906Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3741255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3741610Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3741993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3742366Z return func(*args, **kwargs) 2025-12-04T09:44:32.3742709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3743070Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3743212Z 2025-12-04T09:44:32.3743311Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3743661Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3743962Z res = mod(**inputs) 2025-12-04T09:44:32.3744275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3744618Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3744980Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3745342Z outputs = self.model.decoder( 2025-12-04T09:44:32.3745677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3746017Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3746372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3746740Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3747076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3747423Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3747791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3748145Z return func(*args, **kwargs) 2025-12-04T09:44:32.3748501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3748868Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3749005Z 2025-12-04T09:44:32.3749108Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3749456Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3749768Z res = mod(**inputs) 2025-12-04T09:44:32.3750074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3750420Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3750782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3751145Z outputs = self.model.decoder( 2025-12-04T09:44:32.3751471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3751863Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3752229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3752590Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3752930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3753302Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3753672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3754033Z return func(*args, **kwargs) 2025-12-04T09:44:32.3754398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 287, in forward 2025-12-04T09:44:32.3754817Z hidden_states = (residual + hidden_states).view(hidden_states_shape) 2025-12-04T09:44:32.3755001Z 2025-12-04T09:44:32.3755090Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3755297Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3755508Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3755716Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3755916Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3756123Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3756333Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3756531Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3756765Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3757123Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3757440Z res = mod(**inputs) 2025-12-04T09:44:32.3757757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3758111Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3758494Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3758872Z outputs = self.model.decoder( 2025-12-04T09:44:32.3759231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3759599Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3759992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3760367Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3760723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3761110Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3761563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3761965Z return func(*args, **kwargs) 2025-12-04T09:44:32.3762364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 255, in forward 2025-12-04T09:44:32.3762825Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:32.3763222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3763598Z return func(*args, **kwargs) 2025-12-04T09:44:32.3763966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 179, in forward 2025-12-04T09:44:32.3764388Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:32.3764828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:32.3765304Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:32.3765505Z 2025-12-04T09:44:32.3765649Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3765852Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3766087Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3766439Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3766751Z res = mod(**inputs) 2025-12-04T09:44:32.3767082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3767432Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3767801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3768169Z outputs = self.model.decoder( 2025-12-04T09:44:32.3768507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3768866Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3769233Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3769600Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3769941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3770296Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3770660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3771035Z return func(*args, **kwargs) 2025-12-04T09:44:32.3771394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3771784Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3771935Z 2025-12-04T09:44:32.3772035Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3772392Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3772710Z res = mod(**inputs) 2025-12-04T09:44:32.3773022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3773363Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3773727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3774095Z outputs = self.model.decoder( 2025-12-04T09:44:32.3774434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3774788Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3775162Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3775539Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3775885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3776246Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3776627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3776994Z return func(*args, **kwargs) 2025-12-04T09:44:32.3777359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3777758Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3777908Z 2025-12-04T09:44:32.3778017Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3778368Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3778689Z res = mod(**inputs) 2025-12-04T09:44:32.3779077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3779429Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3779814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3780202Z outputs = self.model.decoder( 2025-12-04T09:44:32.3780548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3780916Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3781288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3781661Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3782002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3782368Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3782749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3783123Z return func(*args, **kwargs) 2025-12-04T09:44:32.3783478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3783862Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3784001Z 2025-12-04T09:44:32.3784112Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3784473Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3784827Z res = mod(**inputs) 2025-12-04T09:44:32.3785155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3785509Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3785884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3786262Z outputs = self.model.decoder( 2025-12-04T09:44:32.3786610Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3786966Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3787338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3787732Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3788093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3788463Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3788847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3789227Z return func(*args, **kwargs) 2025-12-04T09:44:32.3789603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3789990Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3790146Z 2025-12-04T09:44:32.3790235Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3790472Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3790696Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3790922Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3791146Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3791365Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3791565Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3791770Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3792002Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3792353Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3792695Z res = mod(**inputs) 2025-12-04T09:44:32.3793050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3793401Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3793776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3794156Z outputs = self.model.decoder( 2025-12-04T09:44:32.3794517Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3794862Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3795232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3795609Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3795950Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3796299Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3796668Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3797033Z return func(*args, **kwargs) 2025-12-04T09:44:32.3797375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 255, in forward 2025-12-04T09:44:32.3797773Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:32.3798171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3798542Z return func(*args, **kwargs) 2025-12-04T09:44:32.3798895Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 179, in forward 2025-12-04T09:44:32.3799295Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:32.3799745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:32.3800217Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:32.3800407Z 2025-12-04T09:44:32.3800488Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3800702Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3800946Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3801316Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3801718Z res = mod(**inputs) 2025-12-04T09:44:32.3802054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3802406Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3802785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3803176Z outputs = self.model.decoder( 2025-12-04T09:44:32.3803560Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3803949Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3804369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3804794Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3805179Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3805548Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3805937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3806316Z return func(*args, **kwargs) 2025-12-04T09:44:32.3806675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3807133Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3807291Z 2025-12-04T09:44:32.3807393Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3807742Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3808054Z res = mod(**inputs) 2025-12-04T09:44:32.3808397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3808753Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3809122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3809500Z outputs = self.model.decoder( 2025-12-04T09:44:32.3809854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3810215Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3810593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3810976Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3811334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3811698Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3812074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3812445Z return func(*args, **kwargs) 2025-12-04T09:44:32.3812813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3813252Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3813415Z 2025-12-04T09:44:32.3813522Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3813890Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3814217Z res = mod(**inputs) 2025-12-04T09:44:32.3814542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3814902Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3815287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3815666Z outputs = self.model.decoder( 2025-12-04T09:44:32.3816020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3816380Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3816766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3817144Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3817507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3817877Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3818258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3818637Z return func(*args, **kwargs) 2025-12-04T09:44:32.3819012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3819406Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3819548Z 2025-12-04T09:44:32.3819655Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3820022Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3820352Z res = mod(**inputs) 2025-12-04T09:44:32.3820738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3821095Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3821477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3821860Z outputs = self.model.decoder( 2025-12-04T09:44:32.3822204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3822585Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3822963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3823346Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3823696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3824053Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3824432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3824792Z return func(*args, **kwargs) 2025-12-04T09:44:32.3825152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3825537Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3825676Z 2025-12-04T09:44:32.3825790Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3826138Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3826457Z res = mod(**inputs) 2025-12-04T09:44:32.3826780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3827123Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3827495Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3827868Z outputs = self.model.decoder( 2025-12-04T09:44:32.3828212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3828552Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3828922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3829293Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3829638Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3829992Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3830513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3830893Z return func(*args, **kwargs) 2025-12-04T09:44:32.3831257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 287, in forward 2025-12-04T09:44:32.3831695Z hidden_states = (residual + hidden_states).view(hidden_states_shape) 2025-12-04T09:44:32.3831894Z 2025-12-04T09:44:32.3831979Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3832203Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3832428Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3832640Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3832849Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3833050Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3833258Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3833467Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3833696Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3834053Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3834429Z res = mod(**inputs) 2025-12-04T09:44:32.3834820Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3835170Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3835547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3835953Z outputs = self.model.decoder( 2025-12-04T09:44:32.3836291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3836654Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3837029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3837404Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3837748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3838103Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3838471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3838833Z return func(*args, **kwargs) 2025-12-04T09:44:32.3839195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 255, in forward 2025-12-04T09:44:32.3839599Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:32.3840003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3840369Z return func(*args, **kwargs) 2025-12-04T09:44:32.3840736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 179, in forward 2025-12-04T09:44:32.3841147Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:32.3841678Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:32.3842176Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:32.3842389Z 2025-12-04T09:44:32.3842479Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3842716Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3842980Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3843445Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3843772Z res = mod(**inputs) 2025-12-04T09:44:32.3844112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3844463Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3844828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3845204Z outputs = self.model.decoder( 2025-12-04T09:44:32.3845538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3845881Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3846241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3846605Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3846938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3847292Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3847660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3848017Z return func(*args, **kwargs) 2025-12-04T09:44:32.3848405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3848806Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3848953Z 2025-12-04T09:44:32.3849059Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3849400Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3849713Z res = mod(**inputs) 2025-12-04T09:44:32.3850049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3850391Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3850747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3851154Z outputs = self.model.decoder( 2025-12-04T09:44:32.3851503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3851864Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3852241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3852632Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3852991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3853354Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3853757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3854114Z return func(*args, **kwargs) 2025-12-04T09:44:32.3854460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3854848Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3855005Z 2025-12-04T09:44:32.3855112Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3855470Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3855791Z res = mod(**inputs) 2025-12-04T09:44:32.3856126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3856489Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3856875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3857254Z outputs = self.model.decoder( 2025-12-04T09:44:32.3857607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3857970Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3858350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3858737Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3859095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3859467Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3859850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3860233Z return func(*args, **kwargs) 2025-12-04T09:44:32.3860604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3860987Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3861137Z 2025-12-04T09:44:32.3861243Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3861606Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3861932Z res = mod(**inputs) 2025-12-04T09:44:32.3862298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3862654Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3863027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3863398Z outputs = self.model.decoder( 2025-12-04T09:44:32.3863760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3864115Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3864484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3864852Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3865201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3865565Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3865946Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3866313Z return func(*args, **kwargs) 2025-12-04T09:44:32.3866691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3867096Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3867242Z 2025-12-04T09:44:32.3867326Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3867554Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3867778Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3867976Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3868178Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3868380Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3868582Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3868780Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3869012Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3869370Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3869685Z res = mod(**inputs) 2025-12-04T09:44:32.3870009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3870363Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3870734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3871102Z outputs = self.model.decoder( 2025-12-04T09:44:32.3871447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3871800Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3872170Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3872546Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3872893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3873256Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3873631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3874003Z return func(*args, **kwargs) 2025-12-04T09:44:32.3874365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 255, in forward 2025-12-04T09:44:32.3874763Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:32.3875163Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3875589Z return func(*args, **kwargs) 2025-12-04T09:44:32.3876021Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 179, in forward 2025-12-04T09:44:32.3876446Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:32.3876929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:32.3877445Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:32.3877658Z 2025-12-04T09:44:32.3877747Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3877964Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3878212Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3878592Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3878928Z res = mod(**inputs) 2025-12-04T09:44:32.3879272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3879653Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3880051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3880454Z outputs = self.model.decoder( 2025-12-04T09:44:32.3880821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3881197Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3881669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3882083Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3882466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3882883Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3883307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3883711Z return func(*args, **kwargs) 2025-12-04T09:44:32.3884104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3884528Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3884706Z 2025-12-04T09:44:32.3884820Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3885203Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3885543Z res = mod(**inputs) 2025-12-04T09:44:32.3885882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3886259Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3886655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3887054Z outputs = self.model.decoder( 2025-12-04T09:44:32.3887413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3887784Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3888178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3888569Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3888938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3889319Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3889718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3890112Z return func(*args, **kwargs) 2025-12-04T09:44:32.3890562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3891013Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3891176Z 2025-12-04T09:44:32.3891289Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3891676Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3892041Z res = mod(**inputs) 2025-12-04T09:44:32.3892396Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3892777Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3893183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3893585Z outputs = self.model.decoder( 2025-12-04T09:44:32.3893949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3894330Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3894732Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3895141Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3895509Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3895897Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3896310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3896718Z return func(*args, **kwargs) 2025-12-04T09:44:32.3897104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3897520Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3897666Z 2025-12-04T09:44:32.3897787Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3898169Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3898515Z res = mod(**inputs) 2025-12-04T09:44:32.3898865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3899257Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3899652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3900057Z outputs = self.model.decoder( 2025-12-04T09:44:32.3900431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3900812Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3901215Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3901623Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3902004Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3902387Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3902799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3903205Z return func(*args, **kwargs) 2025-12-04T09:44:32.3903590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3904006Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3904161Z 2025-12-04T09:44:32.3904272Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3904667Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3905010Z res = mod(**inputs) 2025-12-04T09:44:32.3905429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3905814Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3906218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3906614Z outputs = self.model.decoder( 2025-12-04T09:44:32.3907008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3907389Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3907786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3908194Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3908576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3908974Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3909381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3909789Z return func(*args, **kwargs) 2025-12-04T09:44:32.3910188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 287, in forward 2025-12-04T09:44:32.3910654Z hidden_states = (residual + hidden_states).view(hidden_states_shape) 2025-12-04T09:44:32.3910865Z 2025-12-04T09:44:32.3910952Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3911188Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3911415Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3911636Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3911861Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3912086Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3912300Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3912527Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3912784Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3913169Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3913542Z res = mod(**inputs) 2025-12-04T09:44:32.3913899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3914288Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3914689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3915121Z outputs = self.model.decoder( 2025-12-04T09:44:32.3915498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3915875Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3916283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3916696Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3917078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3917466Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3917881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3918292Z return func(*args, **kwargs) 2025-12-04T09:44:32.3918696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 255, in forward 2025-12-04T09:44:32.3919141Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:32.3919578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3920010Z return func(*args, **kwargs) 2025-12-04T09:44:32.3920439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 179, in forward 2025-12-04T09:44:32.3920875Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:32.3921377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:32.3922040Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:32.3922248Z 2025-12-04T09:44:32.3922338Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3922580Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3922848Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3923245Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3923593Z res = mod(**inputs) 2025-12-04T09:44:32.3923953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3924394Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3924795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3925214Z outputs = self.model.decoder( 2025-12-04T09:44:32.3925600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3925991Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3926398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3926809Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3927188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3927580Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3927995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3928408Z return func(*args, **kwargs) 2025-12-04T09:44:32.3928806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3929237Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3929413Z 2025-12-04T09:44:32.3929524Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3929912Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3930475Z res = mod(**inputs) 2025-12-04T09:44:32.3930849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3931241Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3931655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3932073Z outputs = self.model.decoder( 2025-12-04T09:44:32.3932463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3932859Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3933264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3933684Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3934069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3934471Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3934880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3935288Z return func(*args, **kwargs) 2025-12-04T09:44:32.3935804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3936244Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3936409Z 2025-12-04T09:44:32.3936521Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3936914Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3937310Z res = mod(**inputs) 2025-12-04T09:44:32.3937668Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3938075Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3938491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3938937Z outputs = self.model.decoder( 2025-12-04T09:44:32.3939307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3939692Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3940096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3940502Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3940886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3941280Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3941690Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3942082Z return func(*args, **kwargs) 2025-12-04T09:44:32.3942472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3942897Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3943050Z 2025-12-04T09:44:32.3943173Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3943558Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3943904Z res = mod(**inputs) 2025-12-04T09:44:32.3944256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3944641Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3945050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3945458Z outputs = self.model.decoder( 2025-12-04T09:44:32.3945835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3946213Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3946619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3947032Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3947401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3947793Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3948207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3948609Z return func(*args, **kwargs) 2025-12-04T09:44:32.3948998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3949414Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3949562Z 2025-12-04T09:44:32.3949657Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3949880Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3950097Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3950357Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3950609Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3950817Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3951039Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3951262Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3951507Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3951896Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3952273Z res = mod(**inputs) 2025-12-04T09:44:32.3952617Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3953000Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3953425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3953830Z outputs = self.model.decoder( 2025-12-04T09:44:32.3954202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3954585Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3954987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3955369Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3955751Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3956145Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3956557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3956964Z return func(*args, **kwargs) 2025-12-04T09:44:32.3957357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 255, in forward 2025-12-04T09:44:32.3957796Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:32.3958228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3958622Z return func(*args, **kwargs) 2025-12-04T09:44:32.3959017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 179, in forward 2025-12-04T09:44:32.3959454Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:32.3959929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:32.3960455Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:32.3960660Z 2025-12-04T09:44:32.3960747Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3960974Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3961223Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3961693Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3962058Z res = mod(**inputs) 2025-12-04T09:44:32.3962416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3962836Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3963252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3963665Z outputs = self.model.decoder( 2025-12-04T09:44:32.3964032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3964417Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3964824Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3965232Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3965684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3966078Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3966500Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3966906Z return func(*args, **kwargs) 2025-12-04T09:44:32.3967327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3967773Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3967939Z 2025-12-04T09:44:32.3968060Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3968452Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3968816Z res = mod(**inputs) 2025-12-04T09:44:32.3969176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3969563Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3969979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3970395Z outputs = self.model.decoder( 2025-12-04T09:44:32.3970775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3971161Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3971573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3971994Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3972370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3972787Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3973237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3973663Z return func(*args, **kwargs) 2025-12-04T09:44:32.3974072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.3974591Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.3974757Z 2025-12-04T09:44:32.3974878Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3975276Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3975637Z res = mod(**inputs) 2025-12-04T09:44:32.3975993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3976389Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3976786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3977196Z outputs = self.model.decoder( 2025-12-04T09:44:32.3977575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3977969Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3978366Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3978782Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3979176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3979577Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3979992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3980397Z return func(*args, **kwargs) 2025-12-04T09:44:32.3980855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3981269Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3981427Z 2025-12-04T09:44:32.3981542Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3981935Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3982304Z res = mod(**inputs) 2025-12-04T09:44:32.3982669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3983065Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3983481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3983954Z outputs = self.model.decoder( 2025-12-04T09:44:32.3984328Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3984724Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3985142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3985559Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3985948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3986354Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3986766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3987182Z return func(*args, **kwargs) 2025-12-04T09:44:32.3987585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.3988012Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.3988165Z 2025-12-04T09:44:32.3988285Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3988684Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3989043Z res = mod(**inputs) 2025-12-04T09:44:32.3989400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3989792Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3990208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3990625Z outputs = self.model.decoder( 2025-12-04T09:44:32.3991001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3991396Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3991813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.3992223Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.3992623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.3993032Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.3993455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.3993859Z return func(*args, **kwargs) 2025-12-04T09:44:32.3994263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 287, in forward 2025-12-04T09:44:32.3994747Z hidden_states = (residual + hidden_states).view(hidden_states_shape) 2025-12-04T09:44:32.3994958Z 2025-12-04T09:44:32.3995054Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3995289Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3995526Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3995783Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3996055Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3996294Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3996528Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3996757Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.3997031Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.3997474Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.3997846Z res = mod(**inputs) 2025-12-04T09:44:32.3998200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.3998595Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.3999003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.3999416Z outputs = self.model.decoder( 2025-12-04T09:44:32.3999794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4000179Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4000587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4000997Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4001380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4001861Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4002281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4002711Z return func(*args, **kwargs) 2025-12-04T09:44:32.4003124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 255, in forward 2025-12-04T09:44:32.4003566Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:32.4003994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4004397Z return func(*args, **kwargs) 2025-12-04T09:44:32.4004791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 179, in forward 2025-12-04T09:44:32.4005227Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:32.4005719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:32.4006249Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:32.4006450Z 2025-12-04T09:44:32.4006546Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4006770Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4007031Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4007425Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4007772Z res = mod(**inputs) 2025-12-04T09:44:32.4008014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4008098Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4008367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4008450Z outputs = self.model.decoder( 2025-12-04T09:44:32.4008688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4008780Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4009042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4009168Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4009438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4009528Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4009794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4009871Z return func(*args, **kwargs) 2025-12-04T09:44:32.4010152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.4010267Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.4010271Z 2025-12-04T09:44:32.4010383Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4010606Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4010677Z res = mod(**inputs) 2025-12-04T09:44:32.4010913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4011003Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4011258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4011347Z outputs = self.model.decoder( 2025-12-04T09:44:32.4011582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4011664Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4011928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4012008Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4012246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4012343Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4012608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4012691Z return func(*args, **kwargs) 2025-12-04T09:44:32.4012947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.4013052Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.4013057Z 2025-12-04T09:44:32.4013173Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4013391Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4013471Z res = mod(**inputs) 2025-12-04T09:44:32.4013710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4013794Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4014078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4014159Z outputs = self.model.decoder( 2025-12-04T09:44:32.4014393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4014481Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4014744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4014835Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4015080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4015169Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4015445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4015523Z return func(*args, **kwargs) 2025-12-04T09:44:32.4015850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.4015952Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.4015956Z 2025-12-04T09:44:32.4016073Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4016307Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4016401Z res = mod(**inputs) 2025-12-04T09:44:32.4016645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4016735Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4017001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4017089Z outputs = self.model.decoder( 2025-12-04T09:44:32.4017330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4017412Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4017684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4017767Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4018013Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4018110Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4018375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4018460Z return func(*args, **kwargs) 2025-12-04T09:44:32.4018725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.4018814Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.4018820Z 2025-12-04T09:44:32.4018918Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4019006Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4019090Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4019182Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4019266Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4019357Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4019441Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4019524Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4019644Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4019862Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4019935Z res = mod(**inputs) 2025-12-04T09:44:32.4020180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4020262Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4020530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4020618Z outputs = self.model.decoder( 2025-12-04T09:44:32.4020855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4020946Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4021218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4021298Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4021549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4021639Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4021910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4022045Z return func(*args, **kwargs) 2025-12-04T09:44:32.4022312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 255, in forward 2025-12-04T09:44:32.4022436Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:32.4022708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4022806Z return func(*args, **kwargs) 2025-12-04T09:44:32.4023079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 179, in forward 2025-12-04T09:44:32.4023189Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:32.4023519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:32.4023669Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:32.4023675Z 2025-12-04T09:44:32.4023765Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4023861Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4023978Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4024201Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4024280Z res = mod(**inputs) 2025-12-04T09:44:32.4024525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4024618Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4024885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4024964Z outputs = self.model.decoder( 2025-12-04T09:44:32.4025217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4025301Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4025578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4025660Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4025905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4026005Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4026275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4026351Z return func(*args, **kwargs) 2025-12-04T09:44:32.4026624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.4026732Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.4026735Z 2025-12-04T09:44:32.4026858Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4027084Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4027155Z res = mod(**inputs) 2025-12-04T09:44:32.4027408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4027492Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4027761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4027849Z outputs = self.model.decoder( 2025-12-04T09:44:32.4028097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4028186Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4028451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4028557Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4028847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4028937Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4029208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4029307Z return func(*args, **kwargs) 2025-12-04T09:44:32.4029575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.4029693Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.4029697Z 2025-12-04T09:44:32.4029809Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4030025Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4030107Z res = mod(**inputs) 2025-12-04T09:44:32.4030551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4030645Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4030914Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4030996Z outputs = self.model.decoder( 2025-12-04T09:44:32.4031247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4031332Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4031598Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4031689Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4031934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4032034Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4032302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4032380Z return func(*args, **kwargs) 2025-12-04T09:44:32.4032657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.4032748Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.4032754Z 2025-12-04T09:44:32.4032877Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4033095Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4033168Z res = mod(**inputs) 2025-12-04T09:44:32.4033413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4033493Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4033759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4033848Z outputs = self.model.decoder( 2025-12-04T09:44:32.4034091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4034180Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4034446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4034528Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4034780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4034867Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4035130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4035213Z return func(*args, **kwargs) 2025-12-04T09:44:32.4035585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.4035684Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.4035689Z 2025-12-04T09:44:32.4035803Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4036020Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4036138Z res = mod(**inputs) 2025-12-04T09:44:32.4036374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4036462Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4036721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4036800Z outputs = self.model.decoder( 2025-12-04T09:44:32.4037047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4037129Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4037391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4037481Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4037727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4037826Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4038098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4038174Z return func(*args, **kwargs) 2025-12-04T09:44:32.4038443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 287, in forward 2025-12-04T09:44:32.4038589Z hidden_states = (residual + hidden_states).view(hidden_states_shape) 2025-12-04T09:44:32.4038594Z 2025-12-04T09:44:32.4038692Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4038778Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4038934Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4039017Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4039101Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4039193Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4039277Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4039359Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4039481Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4039694Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4039766Z res = mod(**inputs) 2025-12-04T09:44:32.4040010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4040092Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4040363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4040441Z outputs = self.model.decoder( 2025-12-04T09:44:32.4040675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4040763Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4041026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4041108Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4041362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4041452Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4041785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4041927Z return func(*args, **kwargs) 2025-12-04T09:44:32.4042198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 255, in forward 2025-12-04T09:44:32.4042328Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:32.4042594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4042707Z return func(*args, **kwargs) 2025-12-04T09:44:32.4042984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 179, in forward 2025-12-04T09:44:32.4043093Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:32.4043416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:32.4043560Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:32.4043565Z 2025-12-04T09:44:32.4043662Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4043749Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4043863Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4044088Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4044158Z res = mod(**inputs) 2025-12-04T09:44:32.4044394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4044485Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4044744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4044831Z outputs = self.model.decoder( 2025-12-04T09:44:32.4045063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4045153Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4045419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4045499Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4045738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4045835Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4046093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4046180Z return func(*args, **kwargs) 2025-12-04T09:44:32.4046439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.4046545Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.4046548Z 2025-12-04T09:44:32.4046670Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4046886Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4046958Z res = mod(**inputs) 2025-12-04T09:44:32.4047202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4047282Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4047556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4047638Z outputs = self.model.decoder( 2025-12-04T09:44:32.4047870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4047958Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4048216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4048329Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4048600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4048689Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4048955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4049049Z return func(*args, **kwargs) 2025-12-04T09:44:32.4049310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 282, in forward 2025-12-04T09:44:32.4049424Z hidden_states = self.activation_fn(hidden_states) 2025-12-04T09:44:32.4049427Z 2025-12-04T09:44:32.4049538Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4049764Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4049834Z res = mod(**inputs) 2025-12-04T09:44:32.4050084Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4050172Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4050431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4050508Z outputs = self.model.decoder( 2025-12-04T09:44:32.4050753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4050832Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4051095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4051174Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4051413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4051508Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4051769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4051851Z return func(*args, **kwargs) 2025-12-04T09:44:32.4052111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.4052201Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.4052205Z 2025-12-04T09:44:32.4052324Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4052539Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4052610Z res = mod(**inputs) 2025-12-04T09:44:32.4052852Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4052932Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4053205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 818, in forward 2025-12-04T09:44:32.4053283Z outputs = self.model.decoder( 2025-12-04T09:44:32.4053518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4053607Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4053869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 648, in forward 2025-12-04T09:44:32.4053948Z layer_outputs = decoder_layer( 2025-12-04T09:44:32.4054195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:32.4054282Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:32.4054552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:32.4054648Z return func(*args, **kwargs) 2025-12-04T09:44:32.4054953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 284, in forward 2025-12-04T09:44:32.4055053Z hidden_states = self.fc2(hidden_states) 2025-12-04T09:44:32.4055057Z 2025-12-04T09:44:32.4055143Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4055238Z cudagraph partition due to non gpu ops 2025-12-04T09:44:32.4055369Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:32.4055581Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:32.4055661Z res = mod(**inputs) 2025-12-04T09:44:32.4055896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/generic.py", line 918, in wrapper 2025-12-04T09:44:32.4055976Z output = func(self, *args, **kwargs) 2025-12-04T09:44:32.4056242Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/opt/modeling_opt.py", line 839, in forward 2025-12-04T09:44:32.4056326Z loss = self.loss_function( 2025-12-04T09:44:32.4056594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/loss/loss_utils.py", line 67, in ForCausalLMLoss 2025-12-04T09:44:32.4056783Z loss = fixed_cross_entropy(logits, shift_labels, num_items_in_batch, ignore_index, **kwargs) 2025-12-04T09:44:32.4057054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/loss/loss_utils.py", line 36, in fixed_cross_entropy 2025-12-04T09:44:32.4057272Z loss = nn.functional.cross_entropy(source, target, ignore_index=ignore_index, reduction=reduction) 2025-12-04T09:44:32.4057276Z 2025-12-04T09:44:43.3810434Z Compilation time (from dynamo_timed): 19.034871875 2025-12-04T09:44:43.4239195Z pass 2025-12-04T09:44:43.4239669Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:44:43.4240611Z TIMING: _recursive_pre_grad_passes:0.03274 _recursive_joint_graph_passes:0.65751 _recursive_post_grad_passes:0.07988 async_compile.wait:1.05228 code_gen:10.34192 inductor_compile:11.96133 backend_compile:16.54348 gc:0.00087 entire_frame_compile:19.03487 total_wall_time:19.03487 2025-12-04T09:44:43.4241815Z STATS: call_* op count: 381 | FakeTensorMode.__torch_dispatch__:19067 | FakeTensor.__torch_dispatch__:3007 | ProxyTorchDispatchMode.__torch_dispatch__:3940 2025-12-04T09:44:43.4242446Z Dynamo produced 1 graphs covering 381 ops with 0 graph breaks (0 unique) 2025-12-04T09:44:46.1982125Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:44:46.1983134Z import pynvml # type: ignore[import] 2025-12-04T09:44:49.6676912Z 2025-12-04T09:44:50.6673897Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:44:50.6674683Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:44:50.6678143Z cpu eval PLBartForCausalLM 2025-12-04T09:44:51.3185438Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:44:51.4962199Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:44:51.6847015Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:44:57.2943160Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2952383Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2957094Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2957431Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2957676Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2957896Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2958105Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2958464Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2959161Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2959404Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2959672Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2959910Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2960143Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2960421Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:57.2960914Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:57.2961284Z res = mod(**inputs) 2025-12-04T09:44:57.2961902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 1680, in forward 2025-12-04T09:44:57.2962378Z outputs = self.model.decoder( 2025-12-04T09:44:57.2962910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 1039, in forward 2025-12-04T09:44:57.2963375Z layer_outputs = decoder_layer( 2025-12-04T09:44:57.2963796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:57.2964222Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:57.2964669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:57.2965103Z return func(*args, **kwargs) 2025-12-04T09:44:57.2965527Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 766, in forward 2025-12-04T09:44:57.2966001Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:57.2966456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:57.2966866Z return func(*args, **kwargs) 2025-12-04T09:44:57.2967289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 443, in forward 2025-12-04T09:44:57.2967825Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:57.2968319Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:57.2968855Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:57.2969065Z 2025-12-04T09:44:57.2969156Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2969392Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2969626Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2969841Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2970067Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2970291Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2970509Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2970731Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2970952Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2971188Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2971403Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2971624Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2971845Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2972097Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:57.2972502Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:57.2972867Z res = mod(**inputs) 2025-12-04T09:44:57.2973271Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 1680, in forward 2025-12-04T09:44:57.2973718Z outputs = self.model.decoder( 2025-12-04T09:44:57.2974152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 1039, in forward 2025-12-04T09:44:57.2974582Z layer_outputs = decoder_layer( 2025-12-04T09:44:57.2975044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:57.2975445Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:57.2975869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:57.2976281Z return func(*args, **kwargs) 2025-12-04T09:44:57.2976711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 766, in forward 2025-12-04T09:44:57.2977167Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:57.2977606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:57.2978000Z return func(*args, **kwargs) 2025-12-04T09:44:57.2978413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 443, in forward 2025-12-04T09:44:57.2978875Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:57.2979385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:57.2979877Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:57.2980065Z 2025-12-04T09:44:57.2980155Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2980369Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2980584Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2980798Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2981001Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2981214Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2981427Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2981629Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2981840Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2982052Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2982261Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2982464Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2982683Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2982938Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:57.2983325Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:57.2983684Z res = mod(**inputs) 2025-12-04T09:44:57.2984092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 1680, in forward 2025-12-04T09:44:57.2984526Z outputs = self.model.decoder( 2025-12-04T09:44:57.2984920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 1039, in forward 2025-12-04T09:44:57.2985372Z layer_outputs = decoder_layer( 2025-12-04T09:44:57.2985754Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:57.2986154Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:57.2986570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:57.2986972Z return func(*args, **kwargs) 2025-12-04T09:44:57.2987407Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 766, in forward 2025-12-04T09:44:57.2987857Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:57.2988275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:57.2988678Z return func(*args, **kwargs) 2025-12-04T09:44:57.2989088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 443, in forward 2025-12-04T09:44:57.2989535Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:57.2990069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:57.2990581Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:57.2990785Z 2025-12-04T09:44:57.2990872Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2991101Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2991365Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2991591Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2991813Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2992037Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2992252Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2992472Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2992692Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2992906Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2993129Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2993353Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2993572Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.2993828Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:57.2994223Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:57.2994589Z res = mod(**inputs) 2025-12-04T09:44:57.2994997Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 1680, in forward 2025-12-04T09:44:57.2995430Z outputs = self.model.decoder( 2025-12-04T09:44:57.2995855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 1039, in forward 2025-12-04T09:44:57.2996274Z layer_outputs = decoder_layer( 2025-12-04T09:44:57.2996655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:57.2997060Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:57.2997473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:57.2997875Z return func(*args, **kwargs) 2025-12-04T09:44:57.2998292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 766, in forward 2025-12-04T09:44:57.2998750Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:57.2999178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:57.2999580Z return func(*args, **kwargs) 2025-12-04T09:44:57.2999996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 443, in forward 2025-12-04T09:44:57.3000450Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:57.3000903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:57.3001395Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:57.3001683Z 2025-12-04T09:44:57.3001786Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3002030Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3002260Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3002496Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3002742Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3002960Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3003185Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3003423Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3003626Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3003840Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3004049Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3004308Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3004522Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3004765Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:57.3005132Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:57.3005454Z res = mod(**inputs) 2025-12-04T09:44:57.3005835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 1680, in forward 2025-12-04T09:44:57.3006261Z outputs = self.model.decoder( 2025-12-04T09:44:57.3006656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 1039, in forward 2025-12-04T09:44:57.3007057Z layer_outputs = decoder_layer( 2025-12-04T09:44:57.3007418Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:57.3007790Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:57.3008178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:57.3008561Z return func(*args, **kwargs) 2025-12-04T09:44:57.3008948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 766, in forward 2025-12-04T09:44:57.3009365Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:57.3009778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:57.3010157Z return func(*args, **kwargs) 2025-12-04T09:44:57.3010541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 443, in forward 2025-12-04T09:44:57.3010959Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:57.3011414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:57.3011902Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:57.3012085Z 2025-12-04T09:44:57.3012174Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3012384Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3012595Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3012807Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3013007Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3013216Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3013429Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3013631Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3013837Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3014047Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3014255Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3014459Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3014670Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3014908Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:57.3015272Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:57.3015604Z res = mod(**inputs) 2025-12-04T09:44:57.3015995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 1680, in forward 2025-12-04T09:44:57.3016386Z outputs = self.model.decoder( 2025-12-04T09:44:57.3016779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 1039, in forward 2025-12-04T09:44:57.3017177Z layer_outputs = decoder_layer( 2025-12-04T09:44:57.3017537Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:44:57.3017903Z return super().__call__(*args, **kwargs) 2025-12-04T09:44:57.3018348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:57.3018731Z return func(*args, **kwargs) 2025-12-04T09:44:57.3019115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 766, in forward 2025-12-04T09:44:57.3019549Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:44:57.3019977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:44:57.3020355Z return func(*args, **kwargs) 2025-12-04T09:44:57.3020737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 443, in forward 2025-12-04T09:44:57.3021162Z attn_output, attn_weights = attention_interface( 2025-12-04T09:44:57.3021622Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:44:57.3022110Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:44:57.3022296Z 2025-12-04T09:44:57.3022377Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3022594Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3022818Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3023015Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3023222Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3023426Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3023636Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3023839Z cudagraph partition due to non gpu ops 2025-12-04T09:44:57.3024080Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:44:57.3024448Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:44:57.3024766Z res = mod(**inputs) 2025-12-04T09:44:57.3025149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/plbart/modeling_plbart.py", line 1702, in forward 2025-12-04T09:44:57.3025622Z loss = loss_fct(logits.view(-1, self.config.vocab_size), labels.view(-1)) 2025-12-04T09:44:57.3025818Z 2025-12-04T09:45:06.4963473Z Compilation time (from dynamo_timed): 13.948648979 2025-12-04T09:45:06.5136835Z pass 2025-12-04T09:45:06.5141277Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:45:06.5146298Z TIMING: _recursive_pre_grad_passes:0.01949 _recursive_joint_graph_passes:0.26699 _recursive_post_grad_passes:0.04105 async_compile.wait:0.87168 code_gen:8.71842 inductor_compile:9.87464 backend_compile:12.44118 gc:0.00054 entire_frame_compile:13.94865 total_wall_time:13.94865 2025-12-04T09:45:06.5148240Z STATS: call_* op count: 182 | FakeTensorMode.__torch_dispatch__:10854 | FakeTensor.__torch_dispatch__:1597 | ProxyTorchDispatchMode.__torch_dispatch__:2157 2025-12-04T09:45:06.5148886Z Dynamo produced 1 graphs covering 182 ops with 0 graph breaks (0 unique) 2025-12-04T09:45:09.0333489Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:45:09.0334374Z import pynvml # type: ignore[import] 2025-12-04T09:45:12.4279013Z 2025-12-04T09:45:15.4830732Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:45:15.4831226Z loading model: 0it [00:03, ?it/s] 2025-12-04T09:45:15.4831495Z cpu eval PegasusForCausalLM 2025-12-04T09:45:15.8202881Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:45:15.9986244Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:45:16.1299807Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:45:25.3633790Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3634788Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3640732Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3640989Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3641848Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3642156Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3643297Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3643514Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3643719Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3643938Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3644147Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3644341Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3644547Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3644750Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3644961Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3645166Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3645372Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3645576Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3645773Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3646016Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3646444Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3646785Z res = mod(**inputs) 2025-12-04T09:45:25.3647198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3647610Z outputs = self.model.decoder( 2025-12-04T09:45:25.3648015Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3648426Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3648803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3649176Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3649575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3649963Z return func(*args, **kwargs) 2025-12-04T09:45:25.3650360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 410, in forward 2025-12-04T09:45:25.3650787Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:45:25.3651192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3651617Z return func(*args, **kwargs) 2025-12-04T09:45:25.3652013Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 255, in forward 2025-12-04T09:45:25.3652450Z attn_output, attn_weights = attention_interface( 2025-12-04T09:45:25.3652904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:45:25.3653401Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:25.3653602Z 2025-12-04T09:45:25.3653688Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3653910Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3654117Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3654329Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3654542Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3654743Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3654953Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3655171Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3655383Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3655731Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3655962Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3656183Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3656406Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3656662Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3657043Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3657377Z res = mod(**inputs) 2025-12-04T09:45:25.3657757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3658173Z outputs = self.model.decoder( 2025-12-04T09:45:25.3658562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3658962Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3659319Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3659688Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3660067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3660442Z return func(*args, **kwargs) 2025-12-04T09:45:25.3660828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 410, in forward 2025-12-04T09:45:25.3661254Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:45:25.3661652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3662023Z return func(*args, **kwargs) 2025-12-04T09:45:25.3662404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 255, in forward 2025-12-04T09:45:25.3662823Z attn_output, attn_weights = attention_interface( 2025-12-04T09:45:25.3663273Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:45:25.3663772Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:25.3663962Z 2025-12-04T09:45:25.3664055Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3664270Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3664487Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3664701Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3664905Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3665119Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3665326Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3665535Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3665736Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3665946Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3666155Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3666357Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3666565Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3666810Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3667175Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3667511Z res = mod(**inputs) 2025-12-04T09:45:25.3667904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3668319Z outputs = self.model.decoder( 2025-12-04T09:45:25.3668718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3669126Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3669526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3669916Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3670307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3670699Z return func(*args, **kwargs) 2025-12-04T09:45:25.3671092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 410, in forward 2025-12-04T09:45:25.3671536Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:45:25.3671946Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3672326Z return func(*args, **kwargs) 2025-12-04T09:45:25.3672713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 255, in forward 2025-12-04T09:45:25.3673144Z attn_output, attn_weights = attention_interface( 2025-12-04T09:45:25.3673607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:45:25.3674098Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:25.3674284Z 2025-12-04T09:45:25.3674367Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3674583Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3674799Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3675008Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3675211Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3675425Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3675666Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3676032Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3676364Z res = mod(**inputs) 2025-12-04T09:45:25.3676757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3677161Z outputs = self.model.decoder( 2025-12-04T09:45:25.3677570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3677981Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3678346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3678716Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3679108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3679493Z return func(*args, **kwargs) 2025-12-04T09:45:25.3679881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 445, in forward 2025-12-04T09:45:25.3680313Z hidden_states = residual + hidden_states 2025-12-04T09:45:25.3680470Z 2025-12-04T09:45:25.3680555Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3680788Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3681010Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3681234Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3681456Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3681805Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3682042Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3682310Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3682721Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3683089Z res = mod(**inputs) 2025-12-04T09:45:25.3683524Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3684003Z outputs = self.model.decoder( 2025-12-04T09:45:25.3684501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3684954Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3685337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3685750Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3686163Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3686573Z return func(*args, **kwargs) 2025-12-04T09:45:25.3686991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 410, in forward 2025-12-04T09:45:25.3687440Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:45:25.3687881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3688290Z return func(*args, **kwargs) 2025-12-04T09:45:25.3688704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 255, in forward 2025-12-04T09:45:25.3689151Z attn_output, attn_weights = attention_interface( 2025-12-04T09:45:25.3689633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:45:25.3690146Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:25.3690340Z 2025-12-04T09:45:25.3690435Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3690656Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3690884Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3691108Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3691324Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3691548Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3691770Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3691988Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3692211Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3692423Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3692632Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3692838Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3693114Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3693357Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3693723Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3694057Z res = mod(**inputs) 2025-12-04T09:45:25.3694441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3694841Z outputs = self.model.decoder( 2025-12-04T09:45:25.3695250Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3695655Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3696017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3696382Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3696775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3697161Z return func(*args, **kwargs) 2025-12-04T09:45:25.3697553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 410, in forward 2025-12-04T09:45:25.3697977Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:45:25.3698388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3698840Z return func(*args, **kwargs) 2025-12-04T09:45:25.3699227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 255, in forward 2025-12-04T09:45:25.3699658Z attn_output, attn_weights = attention_interface( 2025-12-04T09:45:25.3700117Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:45:25.3700625Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:25.3700812Z 2025-12-04T09:45:25.3700893Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3701113Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3701334Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3701540Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3701764Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3701978Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3702221Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3702585Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3702916Z res = mod(**inputs) 2025-12-04T09:45:25.3703301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3703713Z outputs = self.model.decoder( 2025-12-04T09:45:25.3704103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3704499Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3704848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3705205Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3705588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3705962Z return func(*args, **kwargs) 2025-12-04T09:45:25.3706337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 445, in forward 2025-12-04T09:45:25.3706739Z hidden_states = residual + hidden_states 2025-12-04T09:45:25.3706882Z 2025-12-04T09:45:25.3706962Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3707173Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3707374Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3707582Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3707788Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3707985Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3708189Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3708421Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3708779Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3709104Z res = mod(**inputs) 2025-12-04T09:45:25.3709477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3709879Z outputs = self.model.decoder( 2025-12-04T09:45:25.3710266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3710666Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3711020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3711386Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3711762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3712134Z return func(*args, **kwargs) 2025-12-04T09:45:25.3712576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 410, in forward 2025-12-04T09:45:25.3713012Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:45:25.3713416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3713787Z return func(*args, **kwargs) 2025-12-04T09:45:25.3714188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 255, in forward 2025-12-04T09:45:25.3714602Z attn_output, attn_weights = attention_interface( 2025-12-04T09:45:25.3715050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:45:25.3715531Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:25.3715712Z 2025-12-04T09:45:25.3715803Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3716011Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3716219Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3716424Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3716621Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3716827Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3717037Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3717235Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3717443Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3717651Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3717849Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3718054Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3718259Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3718493Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3718845Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3719171Z res = mod(**inputs) 2025-12-04T09:45:25.3719544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3719935Z outputs = self.model.decoder( 2025-12-04T09:45:25.3720324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3720723Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3721071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3721437Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3721919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3722307Z return func(*args, **kwargs) 2025-12-04T09:45:25.3722713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 410, in forward 2025-12-04T09:45:25.3723184Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:45:25.3723629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3724059Z return func(*args, **kwargs) 2025-12-04T09:45:25.3724486Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 255, in forward 2025-12-04T09:45:25.3724912Z attn_output, attn_weights = attention_interface( 2025-12-04T09:45:25.3725391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:45:25.3725936Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:25.3726144Z 2025-12-04T09:45:25.3726231Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3726495Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3726763Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3726984Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3727208Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3727433Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3727680Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3728092Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3728449Z res = mod(**inputs) 2025-12-04T09:45:25.3728857Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3729282Z outputs = self.model.decoder( 2025-12-04T09:45:25.3729761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3730452Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3730839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3731245Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3731668Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3732086Z return func(*args, **kwargs) 2025-12-04T09:45:25.3732505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 445, in forward 2025-12-04T09:45:25.3732926Z hidden_states = residual + hidden_states 2025-12-04T09:45:25.3733065Z 2025-12-04T09:45:25.3733149Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3733358Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3733556Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3733759Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3733965Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3734161Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3734363Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3734591Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3734934Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3735247Z res = mod(**inputs) 2025-12-04T09:45:25.3735622Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3736003Z outputs = self.model.decoder( 2025-12-04T09:45:25.3736372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3736750Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3737097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3737456Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3737826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3738191Z return func(*args, **kwargs) 2025-12-04T09:45:25.3738573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 410, in forward 2025-12-04T09:45:25.3738986Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:45:25.3739380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3739748Z return func(*args, **kwargs) 2025-12-04T09:45:25.3740123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 255, in forward 2025-12-04T09:45:25.3740531Z attn_output, attn_weights = attention_interface( 2025-12-04T09:45:25.3741082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:45:25.3741539Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:25.3741710Z 2025-12-04T09:45:25.3741786Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3741995Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3742205Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3742437Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3742633Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3742835Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3743038Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3743234Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3743439Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3743637Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3743839Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3744038Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3744236Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3744453Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3744795Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3745099Z res = mod(**inputs) 2025-12-04T09:45:25.3745457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3745841Z outputs = self.model.decoder( 2025-12-04T09:45:25.3746239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3746637Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3746963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3747313Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3747677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3748042Z return func(*args, **kwargs) 2025-12-04T09:45:25.3748420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 410, in forward 2025-12-04T09:45:25.3748827Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:45:25.3749216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3749576Z return func(*args, **kwargs) 2025-12-04T09:45:25.3749941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 255, in forward 2025-12-04T09:45:25.3750346Z attn_output, attn_weights = attention_interface( 2025-12-04T09:45:25.3750781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:45:25.3751239Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:25.3751428Z 2025-12-04T09:45:25.3751507Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3751716Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3751926Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3752127Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3752331Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3752546Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3752768Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3753121Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3753435Z res = mod(**inputs) 2025-12-04T09:45:25.3753807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3754268Z outputs = self.model.decoder( 2025-12-04T09:45:25.3754667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3755062Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3755418Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3755797Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3756166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3756532Z return func(*args, **kwargs) 2025-12-04T09:45:25.3756901Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 445, in forward 2025-12-04T09:45:25.3757299Z hidden_states = residual + hidden_states 2025-12-04T09:45:25.3757435Z 2025-12-04T09:45:25.3757519Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3757721Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3757926Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3758126Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3758323Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3758513Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3758715Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3758945Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3759283Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3759595Z res = mod(**inputs) 2025-12-04T09:45:25.3759957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3760337Z outputs = self.model.decoder( 2025-12-04T09:45:25.3760720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3761102Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3761441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3761899Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3762320Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3762724Z return func(*args, **kwargs) 2025-12-04T09:45:25.3763157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 410, in forward 2025-12-04T09:45:25.3763619Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:45:25.3764067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3764474Z return func(*args, **kwargs) 2025-12-04T09:45:25.3764842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 255, in forward 2025-12-04T09:45:25.3765252Z attn_output, attn_weights = attention_interface( 2025-12-04T09:45:25.3765687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:45:25.3766154Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:25.3766334Z 2025-12-04T09:45:25.3766412Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3766619Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3766822Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3767014Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3767215Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3767417Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3767615Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3767863Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3768064Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3768264Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3768457Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3768656Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3768859Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3769100Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3769453Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3769770Z res = mod(**inputs) 2025-12-04T09:45:25.3770139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3770523Z outputs = self.model.decoder( 2025-12-04T09:45:25.3770909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3771300Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3771635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3771991Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3772365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3772735Z return func(*args, **kwargs) 2025-12-04T09:45:25.3773103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 410, in forward 2025-12-04T09:45:25.3773513Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:45:25.3773901Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3774265Z return func(*args, **kwargs) 2025-12-04T09:45:25.3774636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 255, in forward 2025-12-04T09:45:25.3775042Z attn_output, attn_weights = attention_interface( 2025-12-04T09:45:25.3775471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:45:25.3775925Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:25.3776108Z 2025-12-04T09:45:25.3776186Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3776391Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3776596Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3776791Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3776991Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3777191Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3777412Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3777771Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3778084Z res = mod(**inputs) 2025-12-04T09:45:25.3778441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3778830Z outputs = self.model.decoder( 2025-12-04T09:45:25.3779209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3779604Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3779936Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3780291Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3780660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3781034Z return func(*args, **kwargs) 2025-12-04T09:45:25.3781438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 445, in forward 2025-12-04T09:45:25.3781832Z hidden_states = residual + hidden_states 2025-12-04T09:45:25.3781966Z 2025-12-04T09:45:25.3782046Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3782245Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3782462Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3782659Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3782849Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3783052Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3783253Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3783483Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3783831Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3784157Z res = mod(**inputs) 2025-12-04T09:45:25.3784539Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1633, in forward 2025-12-04T09:45:25.3784929Z outputs = self.model.decoder( 2025-12-04T09:45:25.3785329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1120, in forward 2025-12-04T09:45:25.3785717Z layer_outputs = decoder_layer( 2025-12-04T09:45:25.3786062Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:25.3786408Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:25.3786777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3787138Z return func(*args, **kwargs) 2025-12-04T09:45:25.3787510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 410, in forward 2025-12-04T09:45:25.3787934Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:45:25.3788328Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:25.3788691Z return func(*args, **kwargs) 2025-12-04T09:45:25.3789097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 255, in forward 2025-12-04T09:45:25.3789522Z attn_output, attn_weights = attention_interface( 2025-12-04T09:45:25.3789965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward 2025-12-04T09:45:25.3790434Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:25.3790619Z 2025-12-04T09:45:25.3790700Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3790916Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3791130Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3791336Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3791548Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3791761Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3791968Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3792193Z cudagraph partition due to non gpu ops 2025-12-04T09:45:25.3792426Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:25.3792787Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:25.3793105Z res = mod(**inputs) 2025-12-04T09:45:25.3793480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1655, in forward 2025-12-04T09:45:25.3793950Z loss = loss_fct(logits.view(-1, self.config.vocab_size), labels.view(-1)) 2025-12-04T09:45:25.3794145Z 2025-12-04T09:45:34.7935457Z Compilation time (from dynamo_timed): 18.120785094 2025-12-04T09:45:34.7952807Z pass 2025-12-04T09:45:34.7957283Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:45:34.7958111Z TIMING: _recursive_pre_grad_passes:0.03054 _recursive_joint_graph_passes:0.64963 _recursive_post_grad_passes:0.05737 async_compile.wait:0.94382 code_gen:9.60177 inductor_compile:11.16112 backend_compile:15.73468 gc:0.00013 entire_frame_compile:18.12079 total_wall_time:18.12079 2025-12-04T09:45:34.7959062Z STATS: call_* op count: 335 | FakeTensorMode.__torch_dispatch__:20347 | FakeTensor.__torch_dispatch__:2887 | ProxyTorchDispatchMode.__torch_dispatch__:4086 2025-12-04T09:45:34.7959558Z Dynamo produced 1 graphs covering 335 ops with 0 graph breaks (0 unique) 2025-12-04T09:45:37.1081146Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:45:37.1082099Z import pynvml # type: ignore[import] 2025-12-04T09:45:40.2997985Z 2025-12-04T09:45:40.3113493Z loading model: 0it [00:00, ?it/s]If you want to use `RobertaLMHeadModel` as a standalone, add `is_decoder=True.` 2025-12-04T09:45:40.3118161Z WARNING:transformers.models.roberta.modeling_roberta:If you want to use `RobertaLMHeadModel` as a standalone, add `is_decoder=True.` 2025-12-04T09:45:41.3260193Z We strongly recommend passing in an `attention_mask` since your input_ids may be padded. See https://huggingface.co/docs/transformers/troubleshooting#incorrect-output-when-padding-tokens-arent-masked. 2025-12-04T09:45:41.3265994Z You may ignore this warning if your `pad_token_id` (0) is identical to the `bos_token_id` (0), `eos_token_id` (2), or the `sep_token_id` (None), and your input is not padded. 2025-12-04T09:45:41.3268267Z WARNING:transformers.modeling_utils:We strongly recommend passing in an `attention_mask` since your input_ids may be padded. See https://huggingface.co/docs/transformers/troubleshooting#incorrect-output-when-padding-tokens-arent-masked. 2025-12-04T09:45:41.3269323Z You may ignore this warning if your `pad_token_id` (0) is identical to the `bos_token_id` (0), `eos_token_id` (2), or the `sep_token_id` (None), and your input is not padded. 2025-12-04T09:45:41.4746162Z 2025-12-04T09:45:41.4749658Z loading model: 0it [00:01, ?it/s] 2025-12-04T09:45:41.4750236Z cpu eval RobertaForCausalLM 2025-12-04T09:45:42.0232033Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:45:42.2003654Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:45:42.3552623Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:45:51.6251366Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6254460Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6255046Z res = mod(**inputs) 2025-12-04T09:45:51.6260523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6265048Z outputs = self.roberta( 2025-12-04T09:45:51.6270217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 798, in forward 2025-12-04T09:45:51.6275010Z embedding_output = self.embeddings( 2025-12-04T09:45:51.6279417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 87, in forward 2025-12-04T09:45:51.6280103Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length) 2025-12-04T09:45:51.6280733Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 1548, in create_position_ids_from_input_ids 2025-12-04T09:45:51.6281847Z mask = input_ids.ne(padding_idx).int() 2025-12-04T09:45:51.6282027Z 2025-12-04T09:45:51.6282130Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6282368Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6282611Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6282847Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6283127Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6283353Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6283582Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6283808Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6284020Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6284230Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6284443Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6284649Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6284915Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6285325Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6285685Z res = mod(**inputs) 2025-12-04T09:45:51.6286104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6286545Z outputs = self.roberta( 2025-12-04T09:45:51.6286964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 798, in forward 2025-12-04T09:45:51.6287377Z embedding_output = self.embeddings( 2025-12-04T09:45:51.6287795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 87, in forward 2025-12-04T09:45:51.6288331Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length) 2025-12-04T09:45:51.6288927Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 1549, in create_position_ids_from_input_ids 2025-12-04T09:45:51.6289518Z incremental_indices = (torch.cumsum(mask, dim=1).type_as(mask) + past_key_values_length) * mask 2025-12-04T09:45:51.6289772Z 2025-12-04T09:45:51.6289887Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6290264Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6290595Z res = mod(**inputs) 2025-12-04T09:45:51.6290981Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6291408Z outputs = self.roberta( 2025-12-04T09:45:51.6291826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 798, in forward 2025-12-04T09:45:51.6292230Z embedding_output = self.embeddings( 2025-12-04T09:45:51.6292646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 87, in forward 2025-12-04T09:45:51.6293179Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length) 2025-12-04T09:45:51.6293778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 1549, in create_position_ids_from_input_ids 2025-12-04T09:45:51.6294354Z incremental_indices = (torch.cumsum(mask, dim=1).type_as(mask) + past_key_values_length) * mask 2025-12-04T09:45:51.6294601Z 2025-12-04T09:45:51.6294684Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6294906Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6295113Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6295330Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6295541Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6295751Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6295991Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6296318Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6296701Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6297029Z res = mod(**inputs) 2025-12-04T09:45:51.6297425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6297853Z outputs = self.roberta( 2025-12-04T09:45:51.6298235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 862, in forward 2025-12-04T09:45:51.6298649Z encoder_outputs = self.encoder( 2025-12-04T09:45:51.6299058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 606, in forward 2025-12-04T09:45:51.6299466Z layer_outputs = layer_module( 2025-12-04T09:45:51.6299843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:51.6300218Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:51.6300622Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6301012Z return func(*args, **kwargs) 2025-12-04T09:45:51.6301410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 513, in forward 2025-12-04T09:45:51.6301816Z self_attention_outputs = self.attention( 2025-12-04T09:45:51.6302202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6302576Z return func(*args, **kwargs) 2025-12-04T09:45:51.6302957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 440, in forward 2025-12-04T09:45:51.6303343Z self_outputs = self.self( 2025-12-04T09:45:51.6303712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6304086Z return func(*args, **kwargs) 2025-12-04T09:45:51.6304461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 363, in forward 2025-12-04T09:45:51.6304919Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:51.6305118Z 2025-12-04T09:45:51.6305196Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6305403Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6305600Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6305799Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6305997Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6306188Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6306388Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6306588Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6306781Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6306982Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6307182Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6307382Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6307573Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6307804Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6308155Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6308458Z res = mod(**inputs) 2025-12-04T09:45:51.6308819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6309202Z outputs = self.roberta( 2025-12-04T09:45:51.6309562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 862, in forward 2025-12-04T09:45:51.6309996Z encoder_outputs = self.encoder( 2025-12-04T09:45:51.6310376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 606, in forward 2025-12-04T09:45:51.6310756Z layer_outputs = layer_module( 2025-12-04T09:45:51.6311092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:51.6311473Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:51.6311845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6312210Z return func(*args, **kwargs) 2025-12-04T09:45:51.6312576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 513, in forward 2025-12-04T09:45:51.6312970Z self_attention_outputs = self.attention( 2025-12-04T09:45:51.6313350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6313705Z return func(*args, **kwargs) 2025-12-04T09:45:51.6314076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 440, in forward 2025-12-04T09:45:51.6314459Z self_outputs = self.self( 2025-12-04T09:45:51.6314813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6315166Z return func(*args, **kwargs) 2025-12-04T09:45:51.6315536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 363, in forward 2025-12-04T09:45:51.6315971Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:51.6316147Z 2025-12-04T09:45:51.6316232Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6316431Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6316642Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6316854Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6317053Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6317257Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6317462Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6317661Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6317864Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6318070Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6318268Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6318472Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6318676Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6318912Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6319264Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6319591Z res = mod(**inputs) 2025-12-04T09:45:51.6319969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6320356Z outputs = self.roberta( 2025-12-04T09:45:51.6320731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 862, in forward 2025-12-04T09:45:51.6321129Z encoder_outputs = self.encoder( 2025-12-04T09:45:51.6321646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 606, in forward 2025-12-04T09:45:51.6322103Z layer_outputs = layer_module( 2025-12-04T09:45:51.6322508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:51.6322906Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:51.6323318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6323722Z return func(*args, **kwargs) 2025-12-04T09:45:51.6324143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 513, in forward 2025-12-04T09:45:51.6324543Z self_attention_outputs = self.attention( 2025-12-04T09:45:51.6324916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6325302Z return func(*args, **kwargs) 2025-12-04T09:45:51.6325680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 440, in forward 2025-12-04T09:45:51.6326063Z self_outputs = self.self( 2025-12-04T09:45:51.6326428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6326798Z return func(*args, **kwargs) 2025-12-04T09:45:51.6327180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 363, in forward 2025-12-04T09:45:51.6327625Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:51.6327820Z 2025-12-04T09:45:51.6327898Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6328109Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6328315Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6328514Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6328720Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6328924Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6329122Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6329325Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6329532Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6329731Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6329937Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6330299Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6330512Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6330757Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6331125Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6331452Z res = mod(**inputs) 2025-12-04T09:45:51.6331824Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6332222Z outputs = self.roberta( 2025-12-04T09:45:51.6332608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 862, in forward 2025-12-04T09:45:51.6332978Z encoder_outputs = self.encoder( 2025-12-04T09:45:51.6333354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 606, in forward 2025-12-04T09:45:51.6333732Z layer_outputs = layer_module( 2025-12-04T09:45:51.6334069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:51.6334415Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:51.6334835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6335202Z return func(*args, **kwargs) 2025-12-04T09:45:51.6335558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 513, in forward 2025-12-04T09:45:51.6335947Z self_attention_outputs = self.attention( 2025-12-04T09:45:51.6336314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6336668Z return func(*args, **kwargs) 2025-12-04T09:45:51.6337023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 440, in forward 2025-12-04T09:45:51.6337453Z self_outputs = self.self( 2025-12-04T09:45:51.6337857Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6338213Z return func(*args, **kwargs) 2025-12-04T09:45:51.6338566Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 363, in forward 2025-12-04T09:45:51.6339011Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:51.6339181Z 2025-12-04T09:45:51.6339265Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6339458Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6339656Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6339851Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6340045Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6340231Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6340427Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6340626Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6340813Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6341007Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6341198Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6341383Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6341577Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6341800Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6342132Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6342440Z res = mod(**inputs) 2025-12-04T09:45:51.6342790Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6343160Z outputs = self.roberta( 2025-12-04T09:45:51.6343511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 862, in forward 2025-12-04T09:45:51.6343899Z encoder_outputs = self.encoder( 2025-12-04T09:45:51.6344280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 606, in forward 2025-12-04T09:45:51.6344642Z layer_outputs = layer_module( 2025-12-04T09:45:51.6344972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:51.6345320Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:51.6345685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6346027Z return func(*args, **kwargs) 2025-12-04T09:45:51.6346385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 513, in forward 2025-12-04T09:45:51.6346763Z self_attention_outputs = self.attention( 2025-12-04T09:45:51.6347129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6347475Z return func(*args, **kwargs) 2025-12-04T09:45:51.6347832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 440, in forward 2025-12-04T09:45:51.6348199Z self_outputs = self.self( 2025-12-04T09:45:51.6348539Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6348889Z return func(*args, **kwargs) 2025-12-04T09:45:51.6349248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 363, in forward 2025-12-04T09:45:51.6349669Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:51.6349840Z 2025-12-04T09:45:51.6349914Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6350113Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6351075Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6351274Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6351469Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6351661Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6351856Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6352044Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6352259Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6352459Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6352647Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6352840Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6353035Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6353254Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6353600Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6353913Z res = mod(**inputs) 2025-12-04T09:45:51.6354277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6354654Z outputs = self.roberta( 2025-12-04T09:45:51.6355027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 862, in forward 2025-12-04T09:45:51.6355445Z encoder_outputs = self.encoder( 2025-12-04T09:45:51.6355814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 606, in forward 2025-12-04T09:45:51.6356194Z layer_outputs = layer_module( 2025-12-04T09:45:51.6356529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:51.6356879Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:51.6357243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6357616Z return func(*args, **kwargs) 2025-12-04T09:45:51.6357998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 513, in forward 2025-12-04T09:45:51.6358391Z self_attention_outputs = self.attention( 2025-12-04T09:45:51.6358776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6359145Z return func(*args, **kwargs) 2025-12-04T09:45:51.6359526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 440, in forward 2025-12-04T09:45:51.6359910Z self_outputs = self.self( 2025-12-04T09:45:51.6360271Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6360641Z return func(*args, **kwargs) 2025-12-04T09:45:51.6361012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 363, in forward 2025-12-04T09:45:51.6361478Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:51.6361765Z 2025-12-04T09:45:51.6361857Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6362091Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6362315Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6362548Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6362781Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6362982Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6363191Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6363398Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6363605Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6363810Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6364017Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6364224Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6364482Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6364715Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6365071Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6365376Z res = mod(**inputs) 2025-12-04T09:45:51.6365740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6366138Z outputs = self.roberta( 2025-12-04T09:45:51.6366504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 862, in forward 2025-12-04T09:45:51.6366880Z encoder_outputs = self.encoder( 2025-12-04T09:45:51.6367255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 606, in forward 2025-12-04T09:45:51.6367635Z layer_outputs = layer_module( 2025-12-04T09:45:51.6367974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:51.6368332Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:51.6368705Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6369068Z return func(*args, **kwargs) 2025-12-04T09:45:51.6369436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 513, in forward 2025-12-04T09:45:51.6369832Z self_attention_outputs = self.attention( 2025-12-04T09:45:51.6370208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6370569Z return func(*args, **kwargs) 2025-12-04T09:45:51.6370937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 440, in forward 2025-12-04T09:45:51.6371318Z self_outputs = self.self( 2025-12-04T09:45:51.6371676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6372031Z return func(*args, **kwargs) 2025-12-04T09:45:51.6372402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 363, in forward 2025-12-04T09:45:51.6372833Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:51.6373009Z 2025-12-04T09:45:51.6373093Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6373294Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6373498Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6373699Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6373892Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6374094Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6374294Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6374491Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6374690Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6374892Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6375090Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6375283Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6375480Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6375708Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6376060Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6376367Z res = mod(**inputs) 2025-12-04T09:45:51.6376719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6377084Z outputs = self.roberta( 2025-12-04T09:45:51.6377441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 862, in forward 2025-12-04T09:45:51.6377862Z encoder_outputs = self.encoder( 2025-12-04T09:45:51.6378229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 606, in forward 2025-12-04T09:45:51.6378593Z layer_outputs = layer_module( 2025-12-04T09:45:51.6378942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:51.6379304Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:51.6379665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6380032Z return func(*args, **kwargs) 2025-12-04T09:45:51.6380407Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 513, in forward 2025-12-04T09:45:51.6380797Z self_attention_outputs = self.attention( 2025-12-04T09:45:51.6381178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6381541Z return func(*args, **kwargs) 2025-12-04T09:45:51.6381911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 440, in forward 2025-12-04T09:45:51.6382288Z self_outputs = self.self( 2025-12-04T09:45:51.6382645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6383009Z return func(*args, **kwargs) 2025-12-04T09:45:51.6383387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 363, in forward 2025-12-04T09:45:51.6383808Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:51.6383989Z 2025-12-04T09:45:51.6384064Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6384268Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6384461Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6384662Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6384858Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6385049Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6385248Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6385448Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6385646Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6385837Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6386036Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6386242Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6386430Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6386655Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6387001Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6387305Z res = mod(**inputs) 2025-12-04T09:45:51.6387665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6388039Z outputs = self.roberta( 2025-12-04T09:45:51.6388397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 862, in forward 2025-12-04T09:45:51.6388768Z encoder_outputs = self.encoder( 2025-12-04T09:45:51.6389139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 606, in forward 2025-12-04T09:45:51.6389517Z layer_outputs = layer_module( 2025-12-04T09:45:51.6389854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:51.6390207Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:51.6390583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6391008Z return func(*args, **kwargs) 2025-12-04T09:45:51.6391396Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 513, in forward 2025-12-04T09:45:51.6391810Z self_attention_outputs = self.attention( 2025-12-04T09:45:51.6392203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6392617Z return func(*args, **kwargs) 2025-12-04T09:45:51.6393012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 440, in forward 2025-12-04T09:45:51.6393459Z self_outputs = self.self( 2025-12-04T09:45:51.6393834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6394207Z return func(*args, **kwargs) 2025-12-04T09:45:51.6394603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 363, in forward 2025-12-04T09:45:51.6395064Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:51.6395252Z 2025-12-04T09:45:51.6395342Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6395554Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6395774Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6395990Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6396198Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6396412Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6396626Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6396834Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6397054Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6397269Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6397483Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6397691Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6397906Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6398149Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6398512Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6398847Z res = mod(**inputs) 2025-12-04T09:45:51.6399232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6399629Z outputs = self.roberta( 2025-12-04T09:45:51.6400017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 862, in forward 2025-12-04T09:45:51.6400423Z encoder_outputs = self.encoder( 2025-12-04T09:45:51.6400825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 606, in forward 2025-12-04T09:45:51.6401225Z layer_outputs = layer_module( 2025-12-04T09:45:51.6401662Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:51.6402042Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:51.6402434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6402808Z return func(*args, **kwargs) 2025-12-04T09:45:51.6403201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 513, in forward 2025-12-04T09:45:51.6403619Z self_attention_outputs = self.attention( 2025-12-04T09:45:51.6404016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6404388Z return func(*args, **kwargs) 2025-12-04T09:45:51.6404770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 440, in forward 2025-12-04T09:45:51.6405219Z self_outputs = self.self( 2025-12-04T09:45:51.6405586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6405945Z return func(*args, **kwargs) 2025-12-04T09:45:51.6406314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 363, in forward 2025-12-04T09:45:51.6406761Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:51.6406946Z 2025-12-04T09:45:51.6407022Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6407231Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6407434Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6407631Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6407837Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6408038Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6408231Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6408432Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6408630Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6408821Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6409019Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6409217Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6409417Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6409640Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6409988Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6410302Z res = mod(**inputs) 2025-12-04T09:45:51.6410656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6411036Z outputs = self.roberta( 2025-12-04T09:45:51.6411403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 862, in forward 2025-12-04T09:45:51.6411792Z encoder_outputs = self.encoder( 2025-12-04T09:45:51.6412164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 606, in forward 2025-12-04T09:45:51.6412549Z layer_outputs = layer_module( 2025-12-04T09:45:51.6412890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:51.6413278Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:51.6413640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6414004Z return func(*args, **kwargs) 2025-12-04T09:45:51.6414376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 513, in forward 2025-12-04T09:45:51.6414765Z self_attention_outputs = self.attention( 2025-12-04T09:45:51.6415136Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6415499Z return func(*args, **kwargs) 2025-12-04T09:45:51.6415867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 440, in forward 2025-12-04T09:45:51.6416247Z self_outputs = self.self( 2025-12-04T09:45:51.6416596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6416955Z return func(*args, **kwargs) 2025-12-04T09:45:51.6417325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 363, in forward 2025-12-04T09:45:51.6417747Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:51.6417925Z 2025-12-04T09:45:51.6417999Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6418230Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6418459Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6418653Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6418849Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6419049Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6419243Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6419443Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6419664Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6419858Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6420056Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6420257Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6420450Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6420681Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6421028Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6421350Z res = mod(**inputs) 2025-12-04T09:45:51.6421697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward 2025-12-04T09:45:51.6422072Z outputs = self.roberta( 2025-12-04T09:45:51.6422430Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 862, in forward 2025-12-04T09:45:51.6422807Z encoder_outputs = self.encoder( 2025-12-04T09:45:51.6423186Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 606, in forward 2025-12-04T09:45:51.6423566Z layer_outputs = layer_module( 2025-12-04T09:45:51.6423908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:45:51.6424245Z return super().__call__(*args, **kwargs) 2025-12-04T09:45:51.6424605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6424960Z return func(*args, **kwargs) 2025-12-04T09:45:51.6425312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 513, in forward 2025-12-04T09:45:51.6425693Z self_attention_outputs = self.attention( 2025-12-04T09:45:51.6426056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6426407Z return func(*args, **kwargs) 2025-12-04T09:45:51.6426760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 440, in forward 2025-12-04T09:45:51.6427130Z self_outputs = self.self( 2025-12-04T09:45:51.6427474Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:45:51.6427822Z return func(*args, **kwargs) 2025-12-04T09:45:51.6428176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 363, in forward 2025-12-04T09:45:51.6428607Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-12-04T09:45:51.6428780Z 2025-12-04T09:45:51.6428875Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6429069Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6429265Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6429468Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6429664Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6429851Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6430044Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6430420Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6430616Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6430818Z cudagraph partition due to non gpu ops 2025-12-04T09:45:51.6431050Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:45:51.6431555Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:45:51.6431875Z res = mod(**inputs) 2025-12-04T09:45:51.6432242Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/roberta/modeling_roberta.py", line 994, in forward 2025-12-04T09:45:51.6432627Z lm_loss = self.loss_function( 2025-12-04T09:45:51.6432983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/loss/loss_utils.py", line 67, in ForCausalLMLoss 2025-12-04T09:45:51.6433477Z loss = fixed_cross_entropy(logits, shift_labels, num_items_in_batch, ignore_index, **kwargs) 2025-12-04T09:45:51.6433949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/loss/loss_utils.py", line 36, in fixed_cross_entropy 2025-12-04T09:45:51.6434437Z loss = nn.functional.cross_entropy(source, target, ignore_index=ignore_index, reduction=reduction) 2025-12-04T09:45:51.6434688Z 2025-12-04T09:46:01.3375802Z Compilation time (from dynamo_timed): 18.317030569 2025-12-04T09:46:01.3459177Z pass 2025-12-04T09:46:01.3459777Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:46:01.3460797Z TIMING: _recursive_pre_grad_passes:0.02783 _recursive_joint_graph_passes:0.38099 _recursive_post_grad_passes:0.06017 async_compile.wait:0.87402 code_gen:9.22266 inductor_compile:11.09193 backend_compile:15.55414 gc:0.00037 entire_frame_compile:18.31703 total_wall_time:18.31703 2025-12-04T09:46:01.3461940Z STATS: call_* op count: 305 | FakeTensorMode.__torch_dispatch__:20624 | FakeTensor.__torch_dispatch__:3000 | ProxyTorchDispatchMode.__torch_dispatch__:4078 2025-12-04T09:46:01.3462496Z Dynamo produced 1 graphs covering 305 ops with 0 graph breaks (0 unique) 2025-12-04T09:46:04.0729528Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:46:04.0730660Z import pynvml # type: ignore[import] 2025-12-04T09:46:07.5238043Z 2025-12-04T09:46:08.4111917Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:46:08.4112288Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:46:08.4112563Z cpu eval T5ForConditionalGeneration 2025-12-04T09:46:09.4180306Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:46:09.8118551Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:46:10.1885635Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:46:21.1576687Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1577009Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1577218Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1577457Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1577704Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1577931Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1578141Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1578353Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1578560Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1578771Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1578989Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1579194Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1579407Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1579620Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1579831Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1582297Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1582641Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1582873Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1583510Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1583940Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1584214Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1584492Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1584750Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1585032Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1585293Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1585594Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1585820Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1586026Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1586239Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1586464Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1586668Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1586880Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1587100Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1587309Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1587543Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1587758Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1587971Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1588181Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1588393Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1588604Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1588812Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1589028Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1589241Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1589449Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1589703Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1589913Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1590127Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1590359Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1590571Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1590798Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1591018Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1591235Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1591459Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1591724Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:46:21.1592153Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:46:21.1592515Z res = mod(**inputs) 2025-12-04T09:46:21.1592933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1727, in forward 2025-12-04T09:46:21.1593372Z encoder_outputs = self.encoder( 2025-12-04T09:46:21.1593774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1100, in forward 2025-12-04T09:46:21.1594179Z layer_outputs = layer_module( 2025-12-04T09:46:21.1594567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:46:21.1594974Z return super().__call__(*args, **kwargs) 2025-12-04T09:46:21.1595389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:21.1595797Z return func(*args, **kwargs) 2025-12-04T09:46:21.1596201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 687, in forward 2025-12-04T09:46:21.1596620Z self_attention_outputs = self.layer[0]( 2025-12-04T09:46:21.1597035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:21.1597438Z return func(*args, **kwargs) 2025-12-04T09:46:21.1597832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 613, in forward 2025-12-04T09:46:21.1598324Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:46:21.1598582Z 2025-12-04T09:46:21.1598671Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1598903Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1599131Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1599349Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1599578Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1599822Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1600035Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1600261Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1600484Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1600696Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1600918Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1601138Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1601352Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1601744Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1601989Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1602216Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1602433Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1602659Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1602886Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1603103Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1603326Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1603552Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1603769Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1603994Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1604219Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1604434Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1604700Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:46:21.1605102Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:46:21.1605480Z res = mod(**inputs) 2025-12-04T09:46:21.1605875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1764, in forward 2025-12-04T09:46:21.1606292Z decoder_outputs = self.decoder( 2025-12-04T09:46:21.1606700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1100, in forward 2025-12-04T09:46:21.1607102Z layer_outputs = layer_module( 2025-12-04T09:46:21.1607493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:46:21.1607899Z return super().__call__(*args, **kwargs) 2025-12-04T09:46:21.1608313Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:21.1608720Z return func(*args, **kwargs) 2025-12-04T09:46:21.1609110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 687, in forward 2025-12-04T09:46:21.1609532Z self_attention_outputs = self.layer[0]( 2025-12-04T09:46:21.1609944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:21.1610350Z return func(*args, **kwargs) 2025-12-04T09:46:21.1610743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 603, in forward 2025-12-04T09:46:21.1611160Z attention_output = self.SelfAttention( 2025-12-04T09:46:21.1611575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:21.1611986Z return func(*args, **kwargs) 2025-12-04T09:46:21.1612380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 492, in forward 2025-12-04T09:46:21.1612799Z query_states = self.q(hidden_states) 2025-12-04T09:46:21.1612986Z 2025-12-04T09:46:21.1613114Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1613349Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1613558Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1613765Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1613972Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1614185Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1614409Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1614624Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1614832Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1615038Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1615251Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1615467Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1615685Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1615899Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1616154Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:46:21.1616549Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:46:21.1616895Z res = mod(**inputs) 2025-12-04T09:46:21.1617270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1727, in forward 2025-12-04T09:46:21.1617673Z encoder_outputs = self.encoder( 2025-12-04T09:46:21.1618071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1100, in forward 2025-12-04T09:46:21.1618596Z layer_outputs = layer_module( 2025-12-04T09:46:21.1618984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:46:21.1619389Z return super().__call__(*args, **kwargs) 2025-12-04T09:46:21.1619797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:21.1620205Z return func(*args, **kwargs) 2025-12-04T09:46:21.1620579Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 687, in forward 2025-12-04T09:46:21.1620975Z self_attention_outputs = self.layer[0]( 2025-12-04T09:46:21.1621380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:21.1621769Z return func(*args, **kwargs) 2025-12-04T09:46:21.1622144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 613, in forward 2025-12-04T09:46:21.1622584Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:46:21.1622784Z 2025-12-04T09:46:21.1622869Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1623086Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1623298Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1623504Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1623721Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1623931Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1624133Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1624346Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1624553Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1624789Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1625009Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1625217Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1625429Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1625640Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1625851Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1626053Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1626266Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1626477Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1626682Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1626986Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1627199Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1627401Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1627660Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1627861Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1628070Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1628309Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1628511Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1628722Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1628932Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1629135Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1629348Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1629558Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1629764Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1633211Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1633469Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1633680Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1633882Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1634089Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1634297Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1634496Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1634703Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1634908Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1635104Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1635310Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1635515Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1635711Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1635921Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1636126Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1636331Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1636560Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1636803Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:46:21.1637168Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:46:21.1637505Z res = mod(**inputs) 2025-12-04T09:46:21.1637869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1764, in forward 2025-12-04T09:46:21.1638250Z decoder_outputs = self.decoder( 2025-12-04T09:46:21.1638624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1100, in forward 2025-12-04T09:46:21.1639011Z layer_outputs = layer_module( 2025-12-04T09:46:21.1639376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:46:21.1639757Z return super().__call__(*args, **kwargs) 2025-12-04T09:46:21.1640155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:21.1640538Z return func(*args, **kwargs) 2025-12-04T09:46:21.1640952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 687, in forward 2025-12-04T09:46:21.1641368Z self_attention_outputs = self.layer[0]( 2025-12-04T09:46:21.1641879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:21.1642315Z return func(*args, **kwargs) 2025-12-04T09:46:21.1642731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 613, in forward 2025-12-04T09:46:21.1643206Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:46:21.1643412Z 2025-12-04T09:46:21.1643501Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1643718Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1643987Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1644244Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1644458Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1644661Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1644871Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1645086Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1645287Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1645529Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1645747Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1645947Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1646151Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1646359Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1646563Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1646761Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1646966Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1647232Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1647433Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1647635Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1647839Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1648036Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1648244Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1648446Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1648644Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1648850Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1649055Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1649258Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1649459Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1649662Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1649868Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1650066Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1650270Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1650474Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1650703Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:46:21.1651061Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:46:21.1651383Z res = mod(**inputs) 2025-12-04T09:46:21.1651722Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1764, in forward 2025-12-04T09:46:21.1652091Z decoder_outputs = self.decoder( 2025-12-04T09:46:21.1652455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1100, in forward 2025-12-04T09:46:21.1652825Z layer_outputs = layer_module( 2025-12-04T09:46:21.1653166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:46:21.1653536Z return super().__call__(*args, **kwargs) 2025-12-04T09:46:21.1653915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:21.1654282Z return func(*args, **kwargs) 2025-12-04T09:46:21.1654630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 711, in forward 2025-12-04T09:46:21.1655005Z cross_attention_outputs = self.layer[1]( 2025-12-04T09:46:21.1655393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:21.1655758Z return func(*args, **kwargs) 2025-12-04T09:46:21.1656112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 652, in forward 2025-12-04T09:46:21.1656537Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:46:21.1656717Z 2025-12-04T09:46:21.1656808Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1657045Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1657274Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1657486Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1657680Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1657886Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1658094Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1658291Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1658534Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1658734Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1658934Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1659127Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1659327Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1659526Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1659716Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1659916Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1660150Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1660349Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1660554Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1660759Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1660952Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1661154Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1661357Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1661561Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1661755Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1661958Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1662164Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1662357Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1662590Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:46:21.1662951Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:46:21.1663281Z res = mod(**inputs) 2025-12-04T09:46:21.1663628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1764, in forward 2025-12-04T09:46:21.1663999Z decoder_outputs = self.decoder( 2025-12-04T09:46:21.1664358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1100, in forward 2025-12-04T09:46:21.1664716Z layer_outputs = layer_module( 2025-12-04T09:46:21.1665058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:46:21.1665413Z return super().__call__(*args, **kwargs) 2025-12-04T09:46:21.1665777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:21.1666142Z return func(*args, **kwargs) 2025-12-04T09:46:21.1666513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 737, in forward 2025-12-04T09:46:21.1666902Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:46:21.1667289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 344, in forward 2025-12-04T09:46:21.1667723Z hidden_states = hidden_states + self.dropout(forwarded_states) 2025-12-04T09:46:21.1667899Z 2025-12-04T09:46:21.1667982Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1668188Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1668383Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1668583Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1668785Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1668981Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1669182Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1669384Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1669583Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1669820Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1670045Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1670247Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1670448Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1670651Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1670857Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1671050Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1671296Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1671494Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1671688Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1671888Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1672085Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1672277Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1672476Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1672677Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1672899Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1673106Z cudagraph partition due to non gpu ops 2025-12-04T09:46:21.1673338Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:46:21.1673698Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:46:21.1674015Z res = mod(**inputs) 2025-12-04T09:46:21.1674364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1800, in forward 2025-12-04T09:46:21.1674806Z loss = loss_fct(lm_logits.view(-1, lm_logits.size(-1)), labels.view(-1)) 2025-12-04T09:46:21.1674995Z 2025-12-04T09:46:30.8500695Z Compilation time (from dynamo_timed): 19.912631956 2025-12-04T09:46:30.8606102Z pass 2025-12-04T09:46:30.8612116Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:46:30.8617264Z TIMING: _recursive_pre_grad_passes:0.04952 _recursive_joint_graph_passes:0.61862 _recursive_post_grad_passes:0.06405 async_compile.wait:0.77779 code_gen:9.28181 inductor_compile:10.87626 backend_compile:16.98468 gc:0.00014 entire_frame_compile:19.91263 total_wall_time:19.91263 2025-12-04T09:46:30.8618673Z STATS: call_* op count: 824 | FakeTensorMode.__torch_dispatch__:24042 | FakeTensor.__torch_dispatch__:3150 | ProxyTorchDispatchMode.__torch_dispatch__:6109 2025-12-04T09:46:30.8619166Z Dynamo produced 1 graphs covering 824 ops with 0 graph breaks (0 unique) 2025-12-04T09:46:33.2321700Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:46:33.2322504Z import pynvml # type: ignore[import] 2025-12-04T09:46:36.3066176Z 2025-12-04T09:46:37.1756877Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:46:37.1759451Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:46:37.1759917Z cpu eval T5Small 2025-12-04T09:46:38.0517354Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:46:38.4565528Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:46:38.8928276Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:46:49.5037430Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5037732Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5037950Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5038200Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5038404Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5038612Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5038818Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5039022Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5039595Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5039872Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5040083Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5040280Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5040486Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5040692Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5040891Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5041158Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5041365Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5041697Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5041916Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5042123Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5042329Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5042529Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5042735Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5043036Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5043234Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5043433Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5043632Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5043832Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5044044Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5044260Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5044454Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5044656Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5044864Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5045073Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5045321Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5045525Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5045923Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5050640Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5053735Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5054017Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5054238Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5054442Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5054638Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5054843Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5055052Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5055247Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5055449Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5055650Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5055849Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5056043Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5056243Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5056445Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5056637Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5056881Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:46:49.5057261Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:46:49.5057589Z res = mod(**inputs) 2025-12-04T09:46:49.5057964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1727, in forward 2025-12-04T09:46:49.5058349Z encoder_outputs = self.encoder( 2025-12-04T09:46:49.5058721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1100, in forward 2025-12-04T09:46:49.5059080Z layer_outputs = layer_module( 2025-12-04T09:46:49.5059436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:46:49.5059806Z return super().__call__(*args, **kwargs) 2025-12-04T09:46:49.5060190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:49.5061313Z return func(*args, **kwargs) 2025-12-04T09:46:49.5061706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 687, in forward 2025-12-04T09:46:49.5062092Z self_attention_outputs = self.layer[0]( 2025-12-04T09:46:49.5062472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:49.5062900Z return func(*args, **kwargs) 2025-12-04T09:46:49.5063291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 613, in forward 2025-12-04T09:46:49.5063723Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:46:49.5063913Z 2025-12-04T09:46:49.5063994Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5064205Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5064415Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5064648Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5064863Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5065069Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5065265Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5065471Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5065701Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5065906Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5066115Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5066311Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5066515Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5066721Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5066917Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5067124Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5067350Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5067552Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5067757Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5067952Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5068155Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5068357Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5068558Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5068756Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5068958Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5069163Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5069391Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:46:49.5069759Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:46:49.5070114Z res = mod(**inputs) 2025-12-04T09:46:49.5070462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1764, in forward 2025-12-04T09:46:49.5070834Z decoder_outputs = self.decoder( 2025-12-04T09:46:49.5071209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1100, in forward 2025-12-04T09:46:49.5071579Z layer_outputs = layer_module( 2025-12-04T09:46:49.5071923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:46:49.5072290Z return super().__call__(*args, **kwargs) 2025-12-04T09:46:49.5072670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:49.5073034Z return func(*args, **kwargs) 2025-12-04T09:46:49.5073393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 687, in forward 2025-12-04T09:46:49.5073770Z self_attention_outputs = self.layer[0]( 2025-12-04T09:46:49.5074154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:49.5074546Z return func(*args, **kwargs) 2025-12-04T09:46:49.5074922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 603, in forward 2025-12-04T09:46:49.5075308Z attention_output = self.SelfAttention( 2025-12-04T09:46:49.5075694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:49.5076099Z return func(*args, **kwargs) 2025-12-04T09:46:49.5076518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 492, in forward 2025-12-04T09:46:49.5076955Z query_states = self.q(hidden_states) 2025-12-04T09:46:49.5077093Z 2025-12-04T09:46:49.5077172Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5077382Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5077606Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5077800Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5078027Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5078230Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5078428Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5078619Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5078819Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5079019Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5079210Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5079410Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5079609Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5079802Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5080152Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:46:49.5080526Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:46:49.5080853Z res = mod(**inputs) 2025-12-04T09:46:49.5081193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1727, in forward 2025-12-04T09:46:49.5081675Z encoder_outputs = self.encoder( 2025-12-04T09:46:49.5082066Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1100, in forward 2025-12-04T09:46:49.5082462Z layer_outputs = layer_module( 2025-12-04T09:46:49.5082842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:46:49.5083224Z return super().__call__(*args, **kwargs) 2025-12-04T09:46:49.5083612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:49.5083969Z return func(*args, **kwargs) 2025-12-04T09:46:49.5084321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 687, in forward 2025-12-04T09:46:49.5084691Z self_attention_outputs = self.layer[0]( 2025-12-04T09:46:49.5085059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:49.5085427Z return func(*args, **kwargs) 2025-12-04T09:46:49.5085775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 613, in forward 2025-12-04T09:46:49.5086191Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:46:49.5086371Z 2025-12-04T09:46:49.5086450Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5086655Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5086855Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5087048Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5087247Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5087445Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5087643Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5087838Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5088062Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5088278Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5088474Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5088675Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5088876Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5089069Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5089269Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5089484Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5089674Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5089873Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5090070Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5090371Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5090566Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5090765Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5090966Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5091208Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5091414Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5091613Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5091805Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5092003Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5092201Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5092389Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5092592Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5092795Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5092997Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5093192Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5093394Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5093594Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5093785Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5093980Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5094177Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5094373Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5094572Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5094770Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5094960Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5095161Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5095363Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5095561Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5095749Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5095946Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5096141Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5096330Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5096559Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:46:49.5096909Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:46:49.5097219Z res = mod(**inputs) 2025-12-04T09:46:49.5097563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1764, in forward 2025-12-04T09:46:49.5097925Z decoder_outputs = self.decoder( 2025-12-04T09:46:49.5098289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1100, in forward 2025-12-04T09:46:49.5098642Z layer_outputs = layer_module( 2025-12-04T09:46:49.5098982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:46:49.5099337Z return super().__call__(*args, **kwargs) 2025-12-04T09:46:49.5099691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:49.5100040Z return func(*args, **kwargs) 2025-12-04T09:46:49.5100379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 687, in forward 2025-12-04T09:46:49.5100787Z self_attention_outputs = self.layer[0]( 2025-12-04T09:46:49.5101189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:49.5101542Z return func(*args, **kwargs) 2025-12-04T09:46:49.5101884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 613, in forward 2025-12-04T09:46:49.5102297Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:46:49.5102478Z 2025-12-04T09:46:49.5102554Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5102756Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5102952Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5103139Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5103333Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5103529Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5103734Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5103934Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5104129Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5104319Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5104515Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5104711Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5104939Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5105135Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5105323Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5105516Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5105709Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5105895Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5106088Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5106281Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5106468Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5106664Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5106861Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5107049Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5107243Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5107438Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5107635Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5107820Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5108013Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5108205Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5108393Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5108584Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5108777Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5108963Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5109186Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:46:49.5109526Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:46:49.5109830Z res = mod(**inputs) 2025-12-04T09:46:49.5110163Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1764, in forward 2025-12-04T09:46:49.5110514Z decoder_outputs = self.decoder( 2025-12-04T09:46:49.5110858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1100, in forward 2025-12-04T09:46:49.5111198Z layer_outputs = layer_module( 2025-12-04T09:46:49.5111529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:46:49.5111876Z return super().__call__(*args, **kwargs) 2025-12-04T09:46:49.5112235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:49.5112577Z return func(*args, **kwargs) 2025-12-04T09:46:49.5112915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 711, in forward 2025-12-04T09:46:49.5113309Z cross_attention_outputs = self.layer[1]( 2025-12-04T09:46:49.5113671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:49.5114024Z return func(*args, **kwargs) 2025-12-04T09:46:49.5114366Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 652, in forward 2025-12-04T09:46:49.5114785Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-12-04T09:46:49.5114957Z 2025-12-04T09:46:49.5115033Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5115232Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5115430Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5115618Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5115813Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5116030Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5116222Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5116438Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5116632Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5116827Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5117015Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5117208Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5117408Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5117594Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5117788Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5117981Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5118168Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5118364Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5118558Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5118746Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5118944Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5119143Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5119343Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5119537Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5119746Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5119939Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5120127Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5120322Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5120544Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:46:49.5120875Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:46:49.5121183Z res = mod(**inputs) 2025-12-04T09:46:49.5121515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1764, in forward 2025-12-04T09:46:49.5121948Z decoder_outputs = self.decoder( 2025-12-04T09:46:49.5122338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1100, in forward 2025-12-04T09:46:49.5122748Z layer_outputs = layer_module( 2025-12-04T09:46:49.5123118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:46:49.5123516Z return super().__call__(*args, **kwargs) 2025-12-04T09:46:49.5123891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:46:49.5124260Z return func(*args, **kwargs) 2025-12-04T09:46:49.5124604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 737, in forward 2025-12-04T09:46:49.5124967Z hidden_states = self.layer[-1](hidden_states) 2025-12-04T09:46:49.5125353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 344, in forward 2025-12-04T09:46:49.5125778Z hidden_states = hidden_states + self.dropout(forwarded_states) 2025-12-04T09:46:49.5125980Z 2025-12-04T09:46:49.5126081Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5126286Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5126495Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5126701Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5126902Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5127105Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5127326Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5127525Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5127728Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5127934Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5128131Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5128339Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5128544Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5128749Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5128965Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5129174Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5129374Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5129569Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5129771Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5129973Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5130366Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5130582Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5130786Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5130983Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5131190Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5131397Z cudagraph partition due to non gpu ops 2025-12-04T09:46:49.5131635Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:46:49.5131994Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:46:49.5132320Z res = mod(**inputs) 2025-12-04T09:46:49.5132676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1800, in forward 2025-12-04T09:46:49.5133108Z loss = loss_fct(lm_logits.view(-1, lm_logits.size(-1)), labels.view(-1)) 2025-12-04T09:46:49.5133305Z 2025-12-04T09:46:58.1089346Z Compilation time (from dynamo_timed): 18.477584792 2025-12-04T09:46:58.1108011Z pass 2025-12-04T09:46:58.1108844Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:46:58.1109698Z TIMING: _recursive_pre_grad_passes:0.04703 _recursive_joint_graph_passes:0.59914 _recursive_post_grad_passes:0.0623 async_compile.wait:0.00317 code_gen:8.18248 inductor_compile:9.74377 backend_compile:15.64151 gc:0.00064 entire_frame_compile:18.47758 total_wall_time:18.47758 2025-12-04T09:46:58.1110681Z STATS: call_* op count: 824 | FakeTensorMode.__torch_dispatch__:24042 | FakeTensor.__torch_dispatch__:3150 | ProxyTorchDispatchMode.__torch_dispatch__:6109 2025-12-04T09:46:58.1111223Z Dynamo produced 1 graphs covering 824 ops with 0 graph breaks (0 unique) 2025-12-04T09:47:00.4385787Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:47:00.4386588Z import pynvml # type: ignore[import] 2025-12-04T09:47:03.5773779Z 2025-12-04T09:47:05.5269751Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:47:05.5270487Z loading model: 0it [00:01, ?it/s] 2025-12-04T09:47:05.5270818Z cpu eval TrOCRForCausalLM 2025-12-04T09:47:05.6518803Z WARNING:common:fp64 golden ref were not generated for TrOCRForCausalLM. Setting accuracy check to cosine 2025-12-04T09:47:05.6836110Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:47:05.8751830Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:47:06.0280617Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:47:15.2606301Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2606609Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2606832Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2607373Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2607573Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2607783Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2607990Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2608199Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2608398Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2613887Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2614203Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2620186Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2621200Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2621502Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2622188Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2622483Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2622730Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2622999Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2623257Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2623493Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2623728Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2623962Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2624194Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2624432Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2624679Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2624924Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2625148Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2625355Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2625554Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2625759Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2625965Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2626178Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2626397Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2626603Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2626807Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2627022Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2627239Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2627461Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2627661Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2627867Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2628159Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2628365Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2628569Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2628776Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2628972Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2629180Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2629385Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2629603Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2629798Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2629996Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2630395Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2630595Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2630796Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2630998Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2631195Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2631396Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2631784Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2632031Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2632234Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2632430Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2632628Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2632817Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2633015Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2633273Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2633461Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2633657Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2633850Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2634072Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2634282Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2634479Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2634669Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2634919Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2635121Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2635318Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2635509Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2635707Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2635904Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2636095Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2636295Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2636495Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2636686Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2636886Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2637102Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2637293Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2637491Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2637685Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2637912Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2638111Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2638299Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2638498Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2638694Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2638916Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2639110Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2639314Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2639517Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2639713Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2639921Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2640123Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2640325Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2640525Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2640730Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2640937Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2641138Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2641344Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2641723Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2641938Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2642164Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2642387Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2642610Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2642827Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2643033Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2643237Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2643435Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2643642Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2643847Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2644046Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2644292Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2644507Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2644697Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2644892Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2645092Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2645277Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2645472Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2645690Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2645884Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2646073Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2646268Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2646466Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2646653Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2646847Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2647041Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2647247Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2647448Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2647641Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2647829Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2648025Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2648220Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2648413Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2648602Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2648797Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2648993Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2649181Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2649378Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2649573Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2649761Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2649961Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2650158Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2650348Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2650544Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2650738Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2650932Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2651123Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2651316Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2651512Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2651700Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2651895Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2652090Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2652278Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2652472Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2652667Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2652852Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2653053Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2653250Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2653449Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2653650Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2653846Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2654043Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2654231Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2654430Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2654625Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2654820Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2655009Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2655205Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2655402Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2655591Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2655787Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2655985Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2656214Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2656407Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2656602Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2656792Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2656986Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2657180Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2657434Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2657619Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2657812Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2658009Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2658194Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2658388Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2658583Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2658774Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2658989Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2659186Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2659374Z cudagraph partition due to non gpu ops 2025-12-04T09:47:15.2659609Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:15.2659979Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:15.2660304Z res = mod(**inputs) 2025-12-04T09:47:15.2660679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/trocr/modeling_trocr.py", line 849, in forward 2025-12-04T09:47:15.2661135Z loss = loss_fct(logits.view(-1, self.config.vocab_size), labels.view(-1)) 2025-12-04T09:47:15.2661363Z 2025-12-04T09:47:24.3503165Z Compilation time (from dynamo_timed): 17.75841141 2025-12-04T09:47:24.3503658Z pass 2025-12-04T09:47:24.3504549Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:47:24.3505391Z TIMING: _recursive_pre_grad_passes:0.03193 _recursive_joint_graph_passes:0.45722 _recursive_post_grad_passes:0.05909 async_compile.wait:0.80317 code_gen:8.83613 inductor_compile:10.66369 backend_compile:15.39291 gc:0.00027 entire_frame_compile:17.75841 total_wall_time:17.75841 2025-12-04T09:47:24.3506287Z STATS: call_* op count: 445 | FakeTensorMode.__torch_dispatch__:21622 | FakeTensor.__torch_dispatch__:2955 | ProxyTorchDispatchMode.__torch_dispatch__:4644 2025-12-04T09:47:24.3506772Z Dynamo produced 1 graphs covering 445 ops with 0 graph breaks (0 unique) 2025-12-04T09:47:26.8213994Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:47:26.8216317Z import pynvml # type: ignore[import] 2025-12-04T09:47:29.9769080Z 2025-12-04T09:47:35.1321161Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:47:35.1321521Z loading model: 0it [00:05, ?it/s] 2025-12-04T09:47:35.1321909Z cpu eval XGLMForCausalLM 2025-12-04T09:47:35.4722810Z WARNING:common:fp64 golden ref were not generated for XGLMForCausalLM. Setting accuracy check to cosine 2025-12-04T09:47:35.5511683Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:47:36.0331526Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:47:36.5071471Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:47:54.8309100Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8309591Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8309931Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8310714Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8311059Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8312765Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8313133Z res = mod(**inputs) 2025-12-04T09:47:54.8315024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8315591Z outputs = self.model( 2025-12-04T09:47:54.8315993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8316570Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8317061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8317451Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8317928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8318323Z return func(*args, **kwargs) 2025-12-04T09:47:54.8318808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8319276Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8319738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8320146Z return func(*args, **kwargs) 2025-12-04T09:47:54.8320561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8320999Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8321179Z 2025-12-04T09:47:54.8321310Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8321841Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8322270Z res = mod(**inputs) 2025-12-04T09:47:54.8322665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8323088Z outputs = self.model( 2025-12-04T09:47:54.8323490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8323912Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8324295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8324691Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8325114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8325522Z return func(*args, **kwargs) 2025-12-04T09:47:54.8325934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8326445Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8326927Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8327334Z return func(*args, **kwargs) 2025-12-04T09:47:54.8327729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8328212Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8328430Z 2025-12-04T09:47:54.8328521Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8328758Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8329012Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8329405Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8329765Z res = mod(**inputs) 2025-12-04T09:47:54.8330388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8330898Z outputs = self.model( 2025-12-04T09:47:54.8331299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8331719Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8332106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8332529Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8332976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8333327Z return func(*args, **kwargs) 2025-12-04T09:47:54.8333684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8334078Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8334520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8334874Z return func(*args, **kwargs) 2025-12-04T09:47:54.8335233Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8335633Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8335782Z 2025-12-04T09:47:54.8335891Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8336246Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8336568Z res = mod(**inputs) 2025-12-04T09:47:54.8336923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8337290Z outputs = self.model( 2025-12-04T09:47:54.8337701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8338076Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8338417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8338767Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8339139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8339501Z return func(*args, **kwargs) 2025-12-04T09:47:54.8339852Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8340246Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8340635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8340996Z return func(*args, **kwargs) 2025-12-04T09:47:54.8341350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8341775Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8341951Z 2025-12-04T09:47:54.8342039Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8342254Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8342459Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8342660Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8342859Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8343050Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8343255Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8343454Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8343647Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8343848Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8344078Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8344465Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8344784Z res = mod(**inputs) 2025-12-04T09:47:54.8345135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8345504Z outputs = self.model( 2025-12-04T09:47:54.8345850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8346243Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8346591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8346949Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8347316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8347713Z return func(*args, **kwargs) 2025-12-04T09:47:54.8348079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8348473Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8348867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8349232Z return func(*args, **kwargs) 2025-12-04T09:47:54.8349593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8349998Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8350168Z 2025-12-04T09:47:54.8350273Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8350626Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8350939Z res = mod(**inputs) 2025-12-04T09:47:54.8351297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8351668Z outputs = self.model( 2025-12-04T09:47:54.8352021Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8352393Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8352745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8353106Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8353476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8353970Z return func(*args, **kwargs) 2025-12-04T09:47:54.8354334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8354734Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8355121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8355487Z return func(*args, **kwargs) 2025-12-04T09:47:54.8355851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8356291Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8356478Z 2025-12-04T09:47:54.8356558Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8356772Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8357010Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8357356Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8357672Z res = mod(**inputs) 2025-12-04T09:47:54.8358041Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8358424Z outputs = self.model( 2025-12-04T09:47:54.8358774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8359151Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8359499Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8359874Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8360248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8360634Z return func(*args, **kwargs) 2025-12-04T09:47:54.8361017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8361470Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8361982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8362375Z return func(*args, **kwargs) 2025-12-04T09:47:54.8362755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8363173Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8363335Z 2025-12-04T09:47:54.8363441Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8363802Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8364117Z res = mod(**inputs) 2025-12-04T09:47:54.8364477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8364855Z outputs = self.model( 2025-12-04T09:47:54.8365219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8365599Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8365951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8366320Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8366701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8367070Z return func(*args, **kwargs) 2025-12-04T09:47:54.8367437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8367848Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8368240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8368616Z return func(*args, **kwargs) 2025-12-04T09:47:54.8368984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8369419Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8369596Z 2025-12-04T09:47:54.8369677Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8369894Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8370107Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8370313Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8370524Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8370738Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8370955Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8371184Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8371419Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8371637Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8371913Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8372318Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8372642Z res = mod(**inputs) 2025-12-04T09:47:54.8372993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8373372Z outputs = self.model( 2025-12-04T09:47:54.8373750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8374131Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8374474Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8374846Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8375226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8375620Z return func(*args, **kwargs) 2025-12-04T09:47:54.8376006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8376423Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8376835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8377197Z return func(*args, **kwargs) 2025-12-04T09:47:54.8377562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8377985Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8378151Z 2025-12-04T09:47:54.8378264Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8378616Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8378941Z res = mod(**inputs) 2025-12-04T09:47:54.8379297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8379673Z outputs = self.model( 2025-12-04T09:47:54.8380033Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8380413Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8380763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8381119Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8381503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8381879Z return func(*args, **kwargs) 2025-12-04T09:47:54.8382244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8382659Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8383061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8383435Z return func(*args, **kwargs) 2025-12-04T09:47:54.8383796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8384241Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8384437Z 2025-12-04T09:47:54.8384517Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8384730Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8384960Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8385315Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8385634Z res = mod(**inputs) 2025-12-04T09:47:54.8386023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8386403Z outputs = self.model( 2025-12-04T09:47:54.8386763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8387147Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8387502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8387885Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8388268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8388636Z return func(*args, **kwargs) 2025-12-04T09:47:54.8389008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8389436Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8389832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8390188Z return func(*args, **kwargs) 2025-12-04T09:47:54.8390555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8390959Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8391108Z 2025-12-04T09:47:54.8391216Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8391566Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8391894Z res = mod(**inputs) 2025-12-04T09:47:54.8392252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8392629Z outputs = self.model( 2025-12-04T09:47:54.8392998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8393382Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8393746Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8394096Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8394468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8394832Z return func(*args, **kwargs) 2025-12-04T09:47:54.8395190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8395595Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8395990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8396362Z return func(*args, **kwargs) 2025-12-04T09:47:54.8396720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8397151Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8397328Z 2025-12-04T09:47:54.8397414Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8397625Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8397825Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8398030Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8398233Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8398437Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8398673Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8399037Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8399362Z res = mod(**inputs) 2025-12-04T09:47:54.8399774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8400156Z outputs = self.model( 2025-12-04T09:47:54.8400519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8400896Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8401265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8401628Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8402089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8402468Z return func(*args, **kwargs) 2025-12-04T09:47:54.8402838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 369, in forward 2025-12-04T09:47:54.8403260Z hidden_states = residual + hidden_states 2025-12-04T09:47:54.8403406Z 2025-12-04T09:47:54.8403488Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8403706Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8403924Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8404131Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8404378Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8404754Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8405089Z res = mod(**inputs) 2025-12-04T09:47:54.8405439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8405818Z outputs = self.model( 2025-12-04T09:47:54.8406178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8406558Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8406911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8407272Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8407652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8408017Z return func(*args, **kwargs) 2025-12-04T09:47:54.8408387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8408791Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8409200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8409574Z return func(*args, **kwargs) 2025-12-04T09:47:54.8409938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8410362Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8410525Z 2025-12-04T09:47:54.8410629Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8410987Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8411308Z res = mod(**inputs) 2025-12-04T09:47:54.8411666Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8412037Z outputs = self.model( 2025-12-04T09:47:54.8412395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8412777Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8413121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8413565Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8413959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8414338Z return func(*args, **kwargs) 2025-12-04T09:47:54.8414701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8415147Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8415555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8415924Z return func(*args, **kwargs) 2025-12-04T09:47:54.8416301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8416750Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8416938Z 2025-12-04T09:47:54.8417041Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8417252Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8417499Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8417868Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8418199Z res = mod(**inputs) 2025-12-04T09:47:54.8418557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8418929Z outputs = self.model( 2025-12-04T09:47:54.8419284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8419653Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8420002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8420364Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8420747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8421108Z return func(*args, **kwargs) 2025-12-04T09:47:54.8421474Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8421875Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8422267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8422634Z return func(*args, **kwargs) 2025-12-04T09:47:54.8422996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8423397Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8423545Z 2025-12-04T09:47:54.8423647Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8424003Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8424322Z res = mod(**inputs) 2025-12-04T09:47:54.8424663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8425035Z outputs = self.model( 2025-12-04T09:47:54.8425389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8425763Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8426101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8426462Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8426838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8427207Z return func(*args, **kwargs) 2025-12-04T09:47:54.8427589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8427983Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8428370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8428723Z return func(*args, **kwargs) 2025-12-04T09:47:54.8429547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8429967Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8430268Z 2025-12-04T09:47:54.8430362Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8430568Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8430776Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8430981Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8431222Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8431428Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8431631Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8431823Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8432025Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8432233Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8432454Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8432788Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8433100Z res = mod(**inputs) 2025-12-04T09:47:54.8433448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8433801Z outputs = self.model( 2025-12-04T09:47:54.8434151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8434515Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8434846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8435181Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8435542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8435904Z return func(*args, **kwargs) 2025-12-04T09:47:54.8436261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8436647Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8437027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8437378Z return func(*args, **kwargs) 2025-12-04T09:47:54.8437723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8438121Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8438274Z 2025-12-04T09:47:54.8438384Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8438731Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8439038Z res = mod(**inputs) 2025-12-04T09:47:54.8439382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8439749Z outputs = self.model( 2025-12-04T09:47:54.8440092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8440463Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8440804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8441220Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8441574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8441999Z return func(*args, **kwargs) 2025-12-04T09:47:54.8442360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8442784Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8443190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8443598Z return func(*args, **kwargs) 2025-12-04T09:47:54.8443972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8444409Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8444618Z 2025-12-04T09:47:54.8444702Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8444908Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8445130Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8445479Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8445795Z res = mod(**inputs) 2025-12-04T09:47:54.8446141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8446498Z outputs = self.model( 2025-12-04T09:47:54.8446850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8447223Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8447565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8447916Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8448289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8448651Z return func(*args, **kwargs) 2025-12-04T09:47:54.8448999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8449396Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8449786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8450150Z return func(*args, **kwargs) 2025-12-04T09:47:54.8450512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8450919Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8451078Z 2025-12-04T09:47:54.8451188Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8451532Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8451846Z res = mod(**inputs) 2025-12-04T09:47:54.8452188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8452557Z outputs = self.model( 2025-12-04T09:47:54.8452899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8453269Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8453612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8453968Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8454327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8454708Z return func(*args, **kwargs) 2025-12-04T09:47:54.8455075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8455462Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8455852Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8456226Z return func(*args, **kwargs) 2025-12-04T09:47:54.8456578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8456986Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8457165Z 2025-12-04T09:47:54.8457242Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8457450Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8457647Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8457866Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8458068Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8458266Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8458489Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8458836Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8459152Z res = mod(**inputs) 2025-12-04T09:47:54.8459494Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8459865Z outputs = self.model( 2025-12-04T09:47:54.8460217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8460586Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8460920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8461277Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8461648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8462002Z return func(*args, **kwargs) 2025-12-04T09:47:54.8462358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 369, in forward 2025-12-04T09:47:54.8462740Z hidden_states = residual + hidden_states 2025-12-04T09:47:54.8462873Z 2025-12-04T09:47:54.8462955Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8463154Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8463357Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8463560Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8463779Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8464132Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8464448Z res = mod(**inputs) 2025-12-04T09:47:54.8464786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8465154Z outputs = self.model( 2025-12-04T09:47:54.8465509Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8465867Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8466194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8466539Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8466902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8467253Z return func(*args, **kwargs) 2025-12-04T09:47:54.8467594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8468024Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8468405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8468748Z return func(*args, **kwargs) 2025-12-04T09:47:54.8469097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8469507Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8469664Z 2025-12-04T09:47:54.8469771Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8470104Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8470414Z res = mod(**inputs) 2025-12-04T09:47:54.8470753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8471116Z outputs = self.model( 2025-12-04T09:47:54.8471458Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8471819Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8472150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8472488Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8472846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8473198Z return func(*args, **kwargs) 2025-12-04T09:47:54.8473544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8473924Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8474301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8474650Z return func(*args, **kwargs) 2025-12-04T09:47:54.8474988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8475405Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8475587Z 2025-12-04T09:47:54.8475664Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8475861Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8476077Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8476413Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8476720Z res = mod(**inputs) 2025-12-04T09:47:54.8477056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8477415Z outputs = self.model( 2025-12-04T09:47:54.8477758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8478121Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8478452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8478797Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8479162Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8479521Z return func(*args, **kwargs) 2025-12-04T09:47:54.8479881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8480282Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8480672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8481059Z return func(*args, **kwargs) 2025-12-04T09:47:54.8481423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8481892Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8482043Z 2025-12-04T09:47:54.8482155Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8482558Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8482920Z res = mod(**inputs) 2025-12-04T09:47:54.8483320Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8483749Z outputs = self.model( 2025-12-04T09:47:54.8484108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8484509Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8484854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8485199Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8485570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8485932Z return func(*args, **kwargs) 2025-12-04T09:47:54.8486284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8486678Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8487067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8487429Z return func(*args, **kwargs) 2025-12-04T09:47:54.8487781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8488202Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8488372Z 2025-12-04T09:47:54.8488456Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8488663Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8488858Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8489060Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8489259Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8489451Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8489647Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8489844Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8490036Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8490235Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8490463Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8490808Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8491127Z res = mod(**inputs) 2025-12-04T09:47:54.8491470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8491836Z outputs = self.model( 2025-12-04T09:47:54.8492181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8492555Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8492899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8493242Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8493613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8493986Z return func(*args, **kwargs) 2025-12-04T09:47:54.8494359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8494769Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8495160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8495534Z return func(*args, **kwargs) 2025-12-04T09:47:54.8495880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8496287Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8496447Z 2025-12-04T09:47:54.8496546Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8496884Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8497182Z res = mod(**inputs) 2025-12-04T09:47:54.8497535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8497914Z outputs = self.model( 2025-12-04T09:47:54.8498264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8498626Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8498964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8499317Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8499676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8500038Z return func(*args, **kwargs) 2025-12-04T09:47:54.8500395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8500803Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8501198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8501571Z return func(*args, **kwargs) 2025-12-04T09:47:54.8501928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8502355Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8502534Z 2025-12-04T09:47:54.8502609Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8502812Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8503039Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8503373Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8503681Z res = mod(**inputs) 2025-12-04T09:47:54.8504023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8504390Z outputs = self.model( 2025-12-04T09:47:54.8504736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8505104Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8505443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8505805Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8506169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8506528Z return func(*args, **kwargs) 2025-12-04T09:47:54.8506882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8507270Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8507708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8508102Z return func(*args, **kwargs) 2025-12-04T09:47:54.8508458Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8508854Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8509007Z 2025-12-04T09:47:54.8509110Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8509472Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8509779Z res = mod(**inputs) 2025-12-04T09:47:54.8510127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8510494Z outputs = self.model( 2025-12-04T09:47:54.8510841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8511223Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8511576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8511938Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8512307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8512677Z return func(*args, **kwargs) 2025-12-04T09:47:54.8513042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8513445Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8513833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8514209Z return func(*args, **kwargs) 2025-12-04T09:47:54.8514575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8515003Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8515178Z 2025-12-04T09:47:54.8515260Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8515470Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8515683Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8515886Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8516095Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8516302Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8516532Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8516888Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8517213Z res = mod(**inputs) 2025-12-04T09:47:54.8517571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8517944Z outputs = self.model( 2025-12-04T09:47:54.8518302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8518681Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8519024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8519385Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8519763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8520135Z return func(*args, **kwargs) 2025-12-04T09:47:54.8520503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 369, in forward 2025-12-04T09:47:54.8520899Z hidden_states = residual + hidden_states 2025-12-04T09:47:54.8521038Z 2025-12-04T09:47:54.8521128Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8521378Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8521596Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8521893Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8522141Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8522535Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8522920Z res = mod(**inputs) 2025-12-04T09:47:54.8523292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8523675Z outputs = self.model( 2025-12-04T09:47:54.8524058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8524434Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8524795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8525149Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8525521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8525888Z return func(*args, **kwargs) 2025-12-04T09:47:54.8526244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8526645Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8527039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8527403Z return func(*args, **kwargs) 2025-12-04T09:47:54.8527759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8528170Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8528331Z 2025-12-04T09:47:54.8528445Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8528793Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8529101Z res = mod(**inputs) 2025-12-04T09:47:54.8529447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8529813Z outputs = self.model( 2025-12-04T09:47:54.8530291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8530679Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8531029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8531386Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8531755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8532129Z return func(*args, **kwargs) 2025-12-04T09:47:54.8532490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8532881Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8533270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8533632Z return func(*args, **kwargs) 2025-12-04T09:47:54.8533994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8534417Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8534605Z 2025-12-04T09:47:54.8534684Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8534891Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8535169Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8535539Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8535854Z res = mod(**inputs) 2025-12-04T09:47:54.8536200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8536559Z outputs = self.model( 2025-12-04T09:47:54.8536945Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8537323Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8537673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8538029Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8538436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8538815Z return func(*args, **kwargs) 2025-12-04T09:47:54.8539167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8539562Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8539954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8540316Z return func(*args, **kwargs) 2025-12-04T09:47:54.8540670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8541061Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8541207Z 2025-12-04T09:47:54.8541317Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8541660Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8541976Z res = mod(**inputs) 2025-12-04T09:47:54.8542325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8542691Z outputs = self.model( 2025-12-04T09:47:54.8543036Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8543409Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8543801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8544167Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8544550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8544914Z return func(*args, **kwargs) 2025-12-04T09:47:54.8545275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8545667Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8546057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8546417Z return func(*args, **kwargs) 2025-12-04T09:47:54.8546775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8547188Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8547368Z 2025-12-04T09:47:54.8547446Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8547655Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8547851Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8548053Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8548254Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8548453Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8548668Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8548881Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8549090Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8549286Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8549518Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8549879Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8550230Z res = mod(**inputs) 2025-12-04T09:47:54.8550589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8550976Z outputs = self.model( 2025-12-04T09:47:54.8551325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8551690Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8552045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8552398Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8552761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8553124Z return func(*args, **kwargs) 2025-12-04T09:47:54.8553482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8553878Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8554257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8554621Z return func(*args, **kwargs) 2025-12-04T09:47:54.8554993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8555433Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8555597Z 2025-12-04T09:47:54.8555704Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8556063Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8556382Z res = mod(**inputs) 2025-12-04T09:47:54.8556736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8557105Z outputs = self.model( 2025-12-04T09:47:54.8557465Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8557847Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8558193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8558553Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8558939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8559307Z return func(*args, **kwargs) 2025-12-04T09:47:54.8559674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8560079Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8560484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8560848Z return func(*args, **kwargs) 2025-12-04T09:47:54.8561225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8561767Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8561980Z 2025-12-04T09:47:54.8562074Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8562303Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8562605Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8563011Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8563337Z res = mod(**inputs) 2025-12-04T09:47:54.8563713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8564131Z outputs = self.model( 2025-12-04T09:47:54.8564496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8564891Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8565254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8565632Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8566054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8566442Z return func(*args, **kwargs) 2025-12-04T09:47:54.8566827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8567241Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8567662Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8568057Z return func(*args, **kwargs) 2025-12-04T09:47:54.8568450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8568881Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8569041Z 2025-12-04T09:47:54.8569147Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8569519Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8569854Z res = mod(**inputs) 2025-12-04T09:47:54.8570235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8570624Z outputs = self.model( 2025-12-04T09:47:54.8570986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8571362Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8571701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8572062Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8572440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8572814Z return func(*args, **kwargs) 2025-12-04T09:47:54.8573174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8573569Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8573955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8574308Z return func(*args, **kwargs) 2025-12-04T09:47:54.8574663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8575084Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8575258Z 2025-12-04T09:47:54.8575336Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8575544Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8575753Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8575957Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8576151Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8576375Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8576620Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8576965Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8577282Z res = mod(**inputs) 2025-12-04T09:47:54.8577632Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8578008Z outputs = self.model( 2025-12-04T09:47:54.8578358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8578735Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8579073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8579423Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8579814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8580185Z return func(*args, **kwargs) 2025-12-04T09:47:54.8580543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 369, in forward 2025-12-04T09:47:54.8580918Z hidden_states = residual + hidden_states 2025-12-04T09:47:54.8581059Z 2025-12-04T09:47:54.8581138Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8581355Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8581560Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8581768Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8582006Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8582367Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8582685Z res = mod(**inputs) 2025-12-04T09:47:54.8583037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8583477Z outputs = self.model( 2025-12-04T09:47:54.8583834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8584223Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8584575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8584941Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8585315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8585689Z return func(*args, **kwargs) 2025-12-04T09:47:54.8586065Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8586456Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8586855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8587230Z return func(*args, **kwargs) 2025-12-04T09:47:54.8587599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8588015Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8588187Z 2025-12-04T09:47:54.8588292Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8588652Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8588975Z res = mod(**inputs) 2025-12-04T09:47:54.8589325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8589706Z outputs = self.model( 2025-12-04T09:47:54.8590088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8590479Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8590830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8591192Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8591571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8591954Z return func(*args, **kwargs) 2025-12-04T09:47:54.8592319Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8592723Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8593114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8593482Z return func(*args, **kwargs) 2025-12-04T09:47:54.8593866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8594311Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8594495Z 2025-12-04T09:47:54.8594577Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8613833Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8614128Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8614516Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8614854Z res = mod(**inputs) 2025-12-04T09:47:54.8615245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8615633Z outputs = self.model( 2025-12-04T09:47:54.8616008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8616404Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8616762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8617128Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8617525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8617912Z return func(*args, **kwargs) 2025-12-04T09:47:54.8618292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8618702Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8619104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8619475Z return func(*args, **kwargs) 2025-12-04T09:47:54.8619838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8620249Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8620407Z 2025-12-04T09:47:54.8620527Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8620895Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8621234Z res = mod(**inputs) 2025-12-04T09:47:54.8621606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8622008Z outputs = self.model( 2025-12-04T09:47:54.8622368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8622754Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8623102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8623570Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8623946Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8624320Z return func(*args, **kwargs) 2025-12-04T09:47:54.8624687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8625112Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8625512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8625879Z return func(*args, **kwargs) 2025-12-04T09:47:54.8626244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8626666Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8626876Z 2025-12-04T09:47:54.8626962Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8627183Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8627386Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8627592Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8627798Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8628004Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8628202Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8628409Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8628612Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8628808Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8629040Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8629394Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8629707Z res = mod(**inputs) 2025-12-04T09:47:54.8630066Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8630714Z outputs = self.model( 2025-12-04T09:47:54.8631082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8631454Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8631806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8632173Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8632545Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8632921Z return func(*args, **kwargs) 2025-12-04T09:47:54.8633285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8633689Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8634082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8634448Z return func(*args, **kwargs) 2025-12-04T09:47:54.8634812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8635231Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8635398Z 2025-12-04T09:47:54.8635503Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8635862Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8636182Z res = mod(**inputs) 2025-12-04T09:47:54.8636526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8636903Z outputs = self.model( 2025-12-04T09:47:54.8637338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8637715Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8638052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8638407Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8638781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8639172Z return func(*args, **kwargs) 2025-12-04T09:47:54.8639531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8639933Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8640328Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8640753Z return func(*args, **kwargs) 2025-12-04T09:47:54.8641122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8641566Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8641815Z 2025-12-04T09:47:54.8641912Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8642126Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8642372Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8642771Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8643123Z res = mod(**inputs) 2025-12-04T09:47:54.8643491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8643874Z outputs = self.model( 2025-12-04T09:47:54.8644235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8644629Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8644983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8645357Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8645741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8646120Z return func(*args, **kwargs) 2025-12-04T09:47:54.8646491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8646894Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8647297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8647670Z return func(*args, **kwargs) 2025-12-04T09:47:54.8648037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8648473Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8648636Z 2025-12-04T09:47:54.8648742Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8649113Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8649439Z res = mod(**inputs) 2025-12-04T09:47:54.8649819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8650223Z outputs = self.model( 2025-12-04T09:47:54.8650608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8651021Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8651383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8651793Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8652192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8652584Z return func(*args, **kwargs) 2025-12-04T09:47:54.8652974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8653413Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8653831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8654218Z return func(*args, **kwargs) 2025-12-04T09:47:54.8654646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8655061Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8655240Z 2025-12-04T09:47:54.8655319Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8655526Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8655726Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8656112Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8656308Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8656506Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8656725Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8657067Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8657379Z res = mod(**inputs) 2025-12-04T09:47:54.8657722Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8658079Z outputs = self.model( 2025-12-04T09:47:54.8658426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8658792Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8659121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8659477Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8659849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8660211Z return func(*args, **kwargs) 2025-12-04T09:47:54.8660563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 369, in forward 2025-12-04T09:47:54.8660950Z hidden_states = residual + hidden_states 2025-12-04T09:47:54.8661087Z 2025-12-04T09:47:54.8661174Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8661378Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8661589Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8661799Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8662033Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8662386Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8662699Z res = mod(**inputs) 2025-12-04T09:47:54.8663046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8663406Z outputs = self.model( 2025-12-04T09:47:54.8663756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8664131Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8664472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8664824Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8665230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8665596Z return func(*args, **kwargs) 2025-12-04T09:47:54.8665956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8666362Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8666767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8667129Z return func(*args, **kwargs) 2025-12-04T09:47:54.8667475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8667886Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8668046Z 2025-12-04T09:47:54.8668154Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8668522Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8668836Z res = mod(**inputs) 2025-12-04T09:47:54.8669172Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8669539Z outputs = self.model( 2025-12-04T09:47:54.8669877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8670247Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8670587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8670948Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8671309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8671668Z return func(*args, **kwargs) 2025-12-04T09:47:54.8672027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8672417Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8672803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8673170Z return func(*args, **kwargs) 2025-12-04T09:47:54.8673538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8673970Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8674175Z 2025-12-04T09:47:54.8674254Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8674464Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8674688Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8675035Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8675355Z res = mod(**inputs) 2025-12-04T09:47:54.8675699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8676055Z outputs = self.model( 2025-12-04T09:47:54.8676404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8676777Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8677113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8677469Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8677841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8678201Z return func(*args, **kwargs) 2025-12-04T09:47:54.8678566Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8678976Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8679360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8679719Z return func(*args, **kwargs) 2025-12-04T09:47:54.8680067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8680474Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8680620Z 2025-12-04T09:47:54.8680729Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8681071Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8681383Z res = mod(**inputs) 2025-12-04T09:47:54.8681830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8682216Z outputs = self.model( 2025-12-04T09:47:54.8682574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8682967Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8683351Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8683736Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8684130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8684512Z return func(*args, **kwargs) 2025-12-04T09:47:54.8684887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8685299Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8685714Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8686092Z return func(*args, **kwargs) 2025-12-04T09:47:54.8686464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8686903Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8687097Z 2025-12-04T09:47:54.8687179Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8687399Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8687607Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8687821Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8688039Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8688238Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8688322Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8688399Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8688484Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8688562Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8688665Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8688873Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8688939Z res = mod(**inputs) 2025-12-04T09:47:54.8689190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8689267Z outputs = self.model( 2025-12-04T09:47:54.8689513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8689593Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8689815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8689897Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8690187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8690258Z return func(*args, **kwargs) 2025-12-04T09:47:54.8690502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8690611Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8690868Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8690950Z return func(*args, **kwargs) 2025-12-04T09:47:54.8691202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8691320Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8691324Z 2025-12-04T09:47:54.8691465Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8691669Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8691742Z res = mod(**inputs) 2025-12-04T09:47:54.8691987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8692056Z outputs = self.model( 2025-12-04T09:47:54.8692308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8692382Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8692603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8692692Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8692930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8693008Z return func(*args, **kwargs) 2025-12-04T09:47:54.8693248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8693361Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8693599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8693675Z return func(*args, **kwargs) 2025-12-04T09:47:54.8693920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8694053Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8694056Z 2025-12-04T09:47:54.8694144Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8694222Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8694333Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8694533Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8694597Z res = mod(**inputs) 2025-12-04T09:47:54.8694849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8694919Z outputs = self.model( 2025-12-04T09:47:54.8695161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8695244Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8695463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8695548Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8695787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8695857Z return func(*args, **kwargs) 2025-12-04T09:47:54.8696144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8696243Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8696483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8696560Z return func(*args, **kwargs) 2025-12-04T09:47:54.8696825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8696926Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8696930Z 2025-12-04T09:47:54.8697032Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8697229Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8697301Z res = mod(**inputs) 2025-12-04T09:47:54.8697562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8697640Z outputs = self.model( 2025-12-04T09:47:54.8697885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8697958Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8698183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8698263Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8698501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8698578Z return func(*args, **kwargs) 2025-12-04T09:47:54.8698851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8698958Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8699207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8699275Z return func(*args, **kwargs) 2025-12-04T09:47:54.8699529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8699655Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8699661Z 2025-12-04T09:47:54.8699747Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8699823Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8699900Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8699983Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8700057Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8700131Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8700241Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8700476Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8700541Z res = mod(**inputs) 2025-12-04T09:47:54.8700796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8700865Z outputs = self.model( 2025-12-04T09:47:54.8701158Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8701231Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8701450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8701536Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8701777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8701853Z return func(*args, **kwargs) 2025-12-04T09:47:54.8702131Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 369, in forward 2025-12-04T09:47:54.8702214Z hidden_states = residual + hidden_states 2025-12-04T09:47:54.8702217Z 2025-12-04T09:47:54.8702301Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8702376Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8702452Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8702560Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8702663Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8702872Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8702936Z res = mod(**inputs) 2025-12-04T09:47:54.8703183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8703259Z outputs = self.model( 2025-12-04T09:47:54.8703526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8703601Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8703827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8703907Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8704154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8704223Z return func(*args, **kwargs) 2025-12-04T09:47:54.8704467Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8704572Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8704811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8704883Z return func(*args, **kwargs) 2025-12-04T09:47:54.8705135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8705246Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8705250Z 2025-12-04T09:47:54.8705358Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8705556Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8705620Z res = mod(**inputs) 2025-12-04T09:47:54.8705874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8705939Z outputs = self.model( 2025-12-04T09:47:54.8706195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8706267Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8706487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8706574Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8706814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8706882Z return func(*args, **kwargs) 2025-12-04T09:47:54.8707143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8707238Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8707477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8707542Z return func(*args, **kwargs) 2025-12-04T09:47:54.8707781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8707948Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8707954Z 2025-12-04T09:47:54.8708030Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8708104Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8708208Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8708398Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8708503Z res = mod(**inputs) 2025-12-04T09:47:54.8708741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8708807Z outputs = self.model( 2025-12-04T09:47:54.8709051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8709122Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8709353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8709441Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8709685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8709760Z return func(*args, **kwargs) 2025-12-04T09:47:54.8710001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8710098Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8710339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8710405Z return func(*args, **kwargs) 2025-12-04T09:47:54.8710655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8710753Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8710757Z 2025-12-04T09:47:54.8710855Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8711053Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8711117Z res = mod(**inputs) 2025-12-04T09:47:54.8711357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8711431Z outputs = self.model( 2025-12-04T09:47:54.8711671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8711749Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8711963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8712039Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8712284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8712351Z return func(*args, **kwargs) 2025-12-04T09:47:54.8712599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8712694Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8712926Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8713000Z return func(*args, **kwargs) 2025-12-04T09:47:54.8713239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8713359Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8713363Z 2025-12-04T09:47:54.8713445Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8713548Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8713646Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8713722Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8713795Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8713874Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8713947Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8714018Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8714114Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8714186Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8714286Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8714482Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8714545Z res = mod(**inputs) 2025-12-04T09:47:54.8714792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8714874Z outputs = self.model( 2025-12-04T09:47:54.8715113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8715190Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8715406Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8715484Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8715728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8715796Z return func(*args, **kwargs) 2025-12-04T09:47:54.8716048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8716145Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8716383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8716462Z return func(*args, **kwargs) 2025-12-04T09:47:54.8716706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8716827Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8716831Z 2025-12-04T09:47:54.8716933Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8717133Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8717209Z res = mod(**inputs) 2025-12-04T09:47:54.8717467Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8717535Z outputs = self.model( 2025-12-04T09:47:54.8717794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8717871Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8718104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8718185Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8718432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8718513Z return func(*args, **kwargs) 2025-12-04T09:47:54.8718762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8718870Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8719114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8719184Z return func(*args, **kwargs) 2025-12-04T09:47:54.8719445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8719614Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8719618Z 2025-12-04T09:47:54.8719698Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8719785Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8719889Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8720094Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8720176Z res = mod(**inputs) 2025-12-04T09:47:54.8720425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8720500Z outputs = self.model( 2025-12-04T09:47:54.8720748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8720824Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8721074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8721155Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8721406Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8721477Z return func(*args, **kwargs) 2025-12-04T09:47:54.8721809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8721941Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8722197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8722272Z return func(*args, **kwargs) 2025-12-04T09:47:54.8722546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8722653Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8722658Z 2025-12-04T09:47:54.8722768Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8722984Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8723053Z res = mod(**inputs) 2025-12-04T09:47:54.8723328Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8723397Z outputs = self.model( 2025-12-04T09:47:54.8723647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8723728Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8723954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8724042Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8724287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8724359Z return func(*args, **kwargs) 2025-12-04T09:47:54.8724614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8724714Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8724967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8725044Z return func(*args, **kwargs) 2025-12-04T09:47:54.8725292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8725424Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8725428Z 2025-12-04T09:47:54.8725509Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8725612Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8725716Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8725794Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8725872Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8725954Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8726058Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8726284Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8726351Z res = mod(**inputs) 2025-12-04T09:47:54.8726604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8726679Z outputs = self.model( 2025-12-04T09:47:54.8726932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8727020Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8727256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8727336Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8727585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8727656Z return func(*args, **kwargs) 2025-12-04T09:47:54.8727904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 369, in forward 2025-12-04T09:47:54.8727995Z hidden_states = residual + hidden_states 2025-12-04T09:47:54.8727999Z 2025-12-04T09:47:54.8728077Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8728155Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8728240Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8728316Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8728425Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8728626Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8728692Z res = mod(**inputs) 2025-12-04T09:47:54.8728948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8729015Z outputs = self.model( 2025-12-04T09:47:54.8729266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8729347Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8729570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8729656Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8729904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8729978Z return func(*args, **kwargs) 2025-12-04T09:47:54.8730361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8730466Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8730718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8730790Z return func(*args, **kwargs) 2025-12-04T09:47:54.8731041Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8731163Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8731166Z 2025-12-04T09:47:54.8731269Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8731468Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8731585Z res = mod(**inputs) 2025-12-04T09:47:54.8731862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8731941Z outputs = self.model( 2025-12-04T09:47:54.8732198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8732271Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8732527Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8732608Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8732853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8732929Z return func(*args, **kwargs) 2025-12-04T09:47:54.8733218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8733332Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8733582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8733652Z return func(*args, **kwargs) 2025-12-04T09:47:54.8733912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8734057Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8734060Z 2025-12-04T09:47:54.8734143Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8734217Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8734314Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8734514Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8734576Z res = mod(**inputs) 2025-12-04T09:47:54.8734817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8734888Z outputs = self.model( 2025-12-04T09:47:54.8735125Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8735201Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8735420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8735498Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8735743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8735812Z return func(*args, **kwargs) 2025-12-04T09:47:54.8736055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8736163Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8736407Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8736488Z return func(*args, **kwargs) 2025-12-04T09:47:54.8736736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8736831Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8736837Z 2025-12-04T09:47:54.8736945Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8737137Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8737210Z res = mod(**inputs) 2025-12-04T09:47:54.8737458Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8737524Z outputs = self.model( 2025-12-04T09:47:54.8737804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8737877Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8738095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8738181Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8738419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8738507Z return func(*args, **kwargs) 2025-12-04T09:47:54.8738757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8738857Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8739156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8739244Z return func(*args, **kwargs) 2025-12-04T09:47:54.8739481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8739611Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8739614Z 2025-12-04T09:47:54.8739691Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8739772Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8739847Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8739921Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8740001Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8740074Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8740147Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8740228Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8740302Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8740382Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8740485Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8740679Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8740750Z res = mod(**inputs) 2025-12-04T09:47:54.8740995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8741059Z outputs = self.model( 2025-12-04T09:47:54.8741310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8741380Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8741605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8741682Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8741921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8741997Z return func(*args, **kwargs) 2025-12-04T09:47:54.8742237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8742331Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8742574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8742643Z return func(*args, **kwargs) 2025-12-04T09:47:54.8742892Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8742999Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8743003Z 2025-12-04T09:47:54.8743101Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8743300Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8743385Z res = mod(**inputs) 2025-12-04T09:47:54.8743646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8743714Z outputs = self.model( 2025-12-04T09:47:54.8743947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8744042Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8744256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8744334Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8744572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8744638Z return func(*args, **kwargs) 2025-12-04T09:47:54.8744894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8744993Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8745221Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8745295Z return func(*args, **kwargs) 2025-12-04T09:47:54.8745531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8745659Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8745670Z 2025-12-04T09:47:54.8745745Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8745822Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8745928Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8746116Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8746179Z res = mod(**inputs) 2025-12-04T09:47:54.8746427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8746494Z outputs = self.model( 2025-12-04T09:47:54.8746737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8746810Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8747023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8747107Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8747337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8747404Z return func(*args, **kwargs) 2025-12-04T09:47:54.8747645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8747745Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8747983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8748049Z return func(*args, **kwargs) 2025-12-04T09:47:54.8748288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8748389Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8748393Z 2025-12-04T09:47:54.8748491Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8748689Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8748753Z res = mod(**inputs) 2025-12-04T09:47:54.8748991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8749065Z outputs = self.model( 2025-12-04T09:47:54.8749333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8749405Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8749623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8749699Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8749949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8750016Z return func(*args, **kwargs) 2025-12-04T09:47:54.8750253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8750354Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8750598Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8750668Z return func(*args, **kwargs) 2025-12-04T09:47:54.8750916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8751037Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8751040Z 2025-12-04T09:47:54.8751123Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8751201Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8751275Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8751356Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8751431Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8751504Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8751611Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8751804Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8751878Z res = mod(**inputs) 2025-12-04T09:47:54.8752121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8752187Z outputs = self.model( 2025-12-04T09:47:54.8752433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8752503Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8752719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8752803Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8753046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8753121Z return func(*args, **kwargs) 2025-12-04T09:47:54.8753366Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 369, in forward 2025-12-04T09:47:54.8753448Z hidden_states = residual + hidden_states 2025-12-04T09:47:54.8753451Z 2025-12-04T09:47:54.8753535Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8753609Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8753682Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8753763Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8753861Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8754057Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8754121Z res = mod(**inputs) 2025-12-04T09:47:54.8754362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8754438Z outputs = self.model( 2025-12-04T09:47:54.8754678Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8754767Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8755010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8755087Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8755329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8755416Z return func(*args, **kwargs) 2025-12-04T09:47:54.8755703Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8755804Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8756042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8756117Z return func(*args, **kwargs) 2025-12-04T09:47:54.8756380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8756489Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8756493Z 2025-12-04T09:47:54.8756597Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8756785Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8756847Z res = mod(**inputs) 2025-12-04T09:47:54.8757100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8757164Z outputs = self.model( 2025-12-04T09:47:54.8757415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8757486Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8757702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8757791Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8758029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8758095Z return func(*args, **kwargs) 2025-12-04T09:47:54.8758345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8758442Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8758686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8758754Z return func(*args, **kwargs) 2025-12-04T09:47:54.8759007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8759147Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8759153Z 2025-12-04T09:47:54.8759234Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8759320Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8759420Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8759620Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8759691Z res = mod(**inputs) 2025-12-04T09:47:54.8759940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8760009Z outputs = self.model( 2025-12-04T09:47:54.8760265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8760337Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8760567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8760648Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8760920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8761000Z return func(*args, **kwargs) 2025-12-04T09:47:54.8761243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8761341Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8761608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8761742Z return func(*args, **kwargs) 2025-12-04T09:47:54.8762008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8762105Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8762109Z 2025-12-04T09:47:54.8762231Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8762446Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8762515Z res = mod(**inputs) 2025-12-04T09:47:54.8762773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8762844Z outputs = self.model( 2025-12-04T09:47:54.8763099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8763184Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8763411Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8763494Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8763755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8763827Z return func(*args, **kwargs) 2025-12-04T09:47:54.8764081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8764181Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8764417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8764497Z return func(*args, **kwargs) 2025-12-04T09:47:54.8764742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8764872Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8764876Z 2025-12-04T09:47:54.8764953Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8765030Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8765114Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8765192Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8765271Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8765355Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8765431Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8765505Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8765589Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8765664Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8765773Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8765976Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8766041Z res = mod(**inputs) 2025-12-04T09:47:54.8766294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8766362Z outputs = self.model( 2025-12-04T09:47:54.8766610Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8766726Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8766946Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8767034Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8767270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8767354Z return func(*args, **kwargs) 2025-12-04T09:47:54.8767602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8767698Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8767933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8768009Z return func(*args, **kwargs) 2025-12-04T09:47:54.8768270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8768390Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8768394Z 2025-12-04T09:47:54.8768495Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8768695Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8768773Z res = mod(**inputs) 2025-12-04T09:47:54.8769013Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8769085Z outputs = self.model( 2025-12-04T09:47:54.8769325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8769397Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8769618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8769698Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8769932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8770009Z return func(*args, **kwargs) 2025-12-04T09:47:54.8770247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8770351Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8770586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8770655Z return func(*args, **kwargs) 2025-12-04T09:47:54.8770909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8771043Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8771047Z 2025-12-04T09:47:54.8771135Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8771214Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8771316Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8771523Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8771589Z res = mod(**inputs) 2025-12-04T09:47:54.8771843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8771919Z outputs = self.model( 2025-12-04T09:47:54.8772159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8772238Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8772454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8772567Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8772814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8772883Z return func(*args, **kwargs) 2025-12-04T09:47:54.8773126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8773248Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8773493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8773567Z return func(*args, **kwargs) 2025-12-04T09:47:54.8773818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8773914Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8773917Z 2025-12-04T09:47:54.8774042Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8774243Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8774318Z res = mod(**inputs) 2025-12-04T09:47:54.8774565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8774634Z outputs = self.model( 2025-12-04T09:47:54.8774899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8774970Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8775187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8775273Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8775515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8775594Z return func(*args, **kwargs) 2025-12-04T09:47:54.8775839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8775935Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8776184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8776254Z return func(*args, **kwargs) 2025-12-04T09:47:54.8776498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8776626Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8776630Z 2025-12-04T09:47:54.8776707Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8776799Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8776875Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8776953Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8777037Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8777110Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8777210Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8777413Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8777479Z res = mod(**inputs) 2025-12-04T09:47:54.8777734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8777803Z outputs = self.model( 2025-12-04T09:47:54.8778048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8778127Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8778349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8778466Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8778710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8778779Z return func(*args, **kwargs) 2025-12-04T09:47:54.8779030Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 369, in forward 2025-12-04T09:47:54.8779127Z hidden_states = residual + hidden_states 2025-12-04T09:47:54.8779131Z 2025-12-04T09:47:54.8779206Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8779287Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8779361Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8779436Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8779541Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8779751Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8779824Z res = mod(**inputs) 2025-12-04T09:47:54.8780076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8780143Z outputs = self.model( 2025-12-04T09:47:54.8780396Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8780470Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8780705Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8780785Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8781029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8781106Z return func(*args, **kwargs) 2025-12-04T09:47:54.8781355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8781459Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8781702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8781771Z return func(*args, **kwargs) 2025-12-04T09:47:54.8782023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8782134Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8782138Z 2025-12-04T09:47:54.8782238Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8782438Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8782504Z res = mod(**inputs) 2025-12-04T09:47:54.8782750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8782824Z outputs = self.model( 2025-12-04T09:47:54.8783078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8783155Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8783367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8783445Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8783683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8783751Z return func(*args, **kwargs) 2025-12-04T09:47:54.8783998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8784096Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8784357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8784449Z return func(*args, **kwargs) 2025-12-04T09:47:54.8784699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8784829Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8784855Z 2025-12-04T09:47:54.8784934Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8785020Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8785125Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8785316Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8785378Z res = mod(**inputs) 2025-12-04T09:47:54.8785623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8785704Z outputs = self.model( 2025-12-04T09:47:54.8785944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8786022Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8786234Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8786318Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8786550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8786616Z return func(*args, **kwargs) 2025-12-04T09:47:54.8786865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8786960Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8787208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8787278Z return func(*args, **kwargs) 2025-12-04T09:47:54.8787517Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8787617Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8787620Z 2025-12-04T09:47:54.8787717Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8787915Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8787990Z res = mod(**inputs) 2025-12-04T09:47:54.8788235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8788309Z outputs = self.model( 2025-12-04T09:47:54.8788551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8788628Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8788855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8788933Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8789169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8789244Z return func(*args, **kwargs) 2025-12-04T09:47:54.8789485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8789591Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8789828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8789898Z return func(*args, **kwargs) 2025-12-04T09:47:54.8790151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8790318Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8790321Z 2025-12-04T09:47:54.8790409Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8790486Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8790564Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8790650Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8790741Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8790814Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8790894Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8790969Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8791042Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8791123Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8791220Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8791432Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8791502Z res = mod(**inputs) 2025-12-04T09:47:54.8791744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8791821Z outputs = self.model( 2025-12-04T09:47:54.8792073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8792144Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8792362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8792438Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8792676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8792744Z return func(*args, **kwargs) 2025-12-04T09:47:54.8792981Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8793085Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8793315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8793390Z return func(*args, **kwargs) 2025-12-04T09:47:54.8793627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8793736Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8793740Z 2025-12-04T09:47:54.8793848Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8794039Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8794103Z res = mod(**inputs) 2025-12-04T09:47:54.8794349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8794416Z outputs = self.model( 2025-12-04T09:47:54.8794659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8794729Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8794941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8795027Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8795258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8795324Z return func(*args, **kwargs) 2025-12-04T09:47:54.8795569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8795663Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8795933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8796001Z return func(*args, **kwargs) 2025-12-04T09:47:54.8796244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8796383Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8796411Z 2025-12-04T09:47:54.8796488Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8796568Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8796666Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8796857Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8796926Z res = mod(**inputs) 2025-12-04T09:47:54.8797180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8797251Z outputs = self.model( 2025-12-04T09:47:54.8797496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8797566Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8797784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8797861Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8798091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8798167Z return func(*args, **kwargs) 2025-12-04T09:47:54.8798404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8798499Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8798737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8798805Z return func(*args, **kwargs) 2025-12-04T09:47:54.8799048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8799140Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8799144Z 2025-12-04T09:47:54.8799243Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8799443Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8799508Z res = mod(**inputs) 2025-12-04T09:47:54.8799757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8799824Z outputs = self.model( 2025-12-04T09:47:54.8800068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8800151Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8800369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8800446Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8800689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8800759Z return func(*args, **kwargs) 2025-12-04T09:47:54.8801008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8801105Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8801341Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8801419Z return func(*args, **kwargs) 2025-12-04T09:47:54.8801864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8802066Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8802070Z 2025-12-04T09:47:54.8802157Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8802242Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8802337Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8802440Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8802523Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8802615Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8802727Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8802943Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8803019Z res = mod(**inputs) 2025-12-04T09:47:54.8803289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8803370Z outputs = self.model( 2025-12-04T09:47:54.8803612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8803685Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8803915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8803996Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8804249Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8804318Z return func(*args, **kwargs) 2025-12-04T09:47:54.8804556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 369, in forward 2025-12-04T09:47:54.8804643Z hidden_states = residual + hidden_states 2025-12-04T09:47:54.8804646Z 2025-12-04T09:47:54.8804727Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8804805Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8804889Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8804962Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8805071Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8805268Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8805334Z res = mod(**inputs) 2025-12-04T09:47:54.8805584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8805651Z outputs = self.model( 2025-12-04T09:47:54.8805890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8805969Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8806188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8806276Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8806510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8806578Z return func(*args, **kwargs) 2025-12-04T09:47:54.8806825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8806923Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8807155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8807234Z return func(*args, **kwargs) 2025-12-04T09:47:54.8807474Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8807591Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8807613Z 2025-12-04T09:47:54.8807729Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8807926Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8808002Z res = mod(**inputs) 2025-12-04T09:47:54.8808243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8808336Z outputs = self.model( 2025-12-04T09:47:54.8808581Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8808654Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8808880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8808961Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8809213Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8809295Z return func(*args, **kwargs) 2025-12-04T09:47:54.8809544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8809648Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8809880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8809948Z return func(*args, **kwargs) 2025-12-04T09:47:54.8810197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8810332Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8810335Z 2025-12-04T09:47:54.8810419Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8810499Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8810605Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8810808Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8810874Z res = mod(**inputs) 2025-12-04T09:47:54.8811124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8811202Z outputs = self.model( 2025-12-04T09:47:54.8811449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8811537Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8811754Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8811833Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8812078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8812150Z return func(*args, **kwargs) 2025-12-04T09:47:54.8812391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8812496Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8812731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8812806Z return func(*args, **kwargs) 2025-12-04T09:47:54.8813047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8813141Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8813145Z 2025-12-04T09:47:54.8813251Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8813447Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8813530Z res = mod(**inputs) 2025-12-04T09:47:54.8814152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8814229Z outputs = self.model( 2025-12-04T09:47:54.8814478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8814567Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8814784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8814871Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8815113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8815193Z return func(*args, **kwargs) 2025-12-04T09:47:54.8815452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8815554Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8815799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8815866Z return func(*args, **kwargs) 2025-12-04T09:47:54.8816110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8816243Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8816247Z 2025-12-04T09:47:54.8816325Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8816411Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8816488Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8816574Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8816655Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8816730Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8816806Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8816888Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8816962Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8817035Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8817141Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8817355Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8817428Z res = mod(**inputs) 2025-12-04T09:47:54.8817664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8817730Z outputs = self.model( 2025-12-04T09:47:54.8817976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8818046Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8818262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8818347Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8818576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8818649Z return func(*args, **kwargs) 2025-12-04T09:47:54.8818888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8818984Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8819223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8819289Z return func(*args, **kwargs) 2025-12-04T09:47:54.8819531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8819660Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8819678Z 2025-12-04T09:47:54.8819777Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8819974Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8820039Z res = mod(**inputs) 2025-12-04T09:47:54.8820278Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8820367Z outputs = self.model( 2025-12-04T09:47:54.8820606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8820685Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8820903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8820979Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8821243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8821313Z return func(*args, **kwargs) 2025-12-04T09:47:54.8821561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8821656Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8821886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8821962Z return func(*args, **kwargs) 2025-12-04T09:47:54.8822199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8822327Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8822331Z 2025-12-04T09:47:54.8822416Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8822494Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8822598Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8822788Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8822852Z res = mod(**inputs) 2025-12-04T09:47:54.8823097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8823164Z outputs = self.model( 2025-12-04T09:47:54.8823401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8823478Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8823688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8823773Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8824008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8824075Z return func(*args, **kwargs) 2025-12-04T09:47:54.8824317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8824413Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8824649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8824717Z return func(*args, **kwargs) 2025-12-04T09:47:54.8824952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8825050Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8825054Z 2025-12-04T09:47:54.8825150Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8825338Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8825442Z res = mod(**inputs) 2025-12-04T09:47:54.8825680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8825754Z outputs = self.model( 2025-12-04T09:47:54.8825991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8826076Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8826296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8826373Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8826603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8826675Z return func(*args, **kwargs) 2025-12-04T09:47:54.8826929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8827033Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8827262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8827329Z return func(*args, **kwargs) 2025-12-04T09:47:54.8827571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8827690Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8827694Z 2025-12-04T09:47:54.8827776Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8827851Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8827925Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8828007Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8828080Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8828155Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8828262Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8828449Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8828513Z res = mod(**inputs) 2025-12-04T09:47:54.8828755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8828822Z outputs = self.model( 2025-12-04T09:47:54.8829069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8829139Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8829348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8829431Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8829662Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8829736Z return func(*args, **kwargs) 2025-12-04T09:47:54.8829969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 369, in forward 2025-12-04T09:47:54.8830049Z hidden_states = residual + hidden_states 2025-12-04T09:47:54.8830052Z 2025-12-04T09:47:54.8830317Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8830403Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8830480Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8830564Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8830662Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8830864Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8830932Z res = mod(**inputs) 2025-12-04T09:47:54.8831222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8831320Z outputs = self.model( 2025-12-04T09:47:54.8831554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8831624Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8831843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8831944Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8832182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8832249Z return func(*args, **kwargs) 2025-12-04T09:47:54.8832486Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8832641Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8832878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8832946Z return func(*args, **kwargs) 2025-12-04T09:47:54.8833188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 157, in forward 2025-12-04T09:47:54.8833297Z query_states = self.q_proj(hidden_states) * self.scaling 2025-12-04T09:47:54.8833302Z 2025-12-04T09:47:54.8833408Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8833595Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8833659Z res = mod(**inputs) 2025-12-04T09:47:54.8833902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8833967Z outputs = self.model( 2025-12-04T09:47:54.8834223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8834294Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8834506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8834590Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8834824Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8834893Z return func(*args, **kwargs) 2025-12-04T09:47:54.8835139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8835235Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8835474Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8835544Z return func(*args, **kwargs) 2025-12-04T09:47:54.8835784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 199, in forward 2025-12-04T09:47:54.8835921Z attn_weights = torch.bmm(query_states, key_states.transpose(1, 2)) 2025-12-04T09:47:54.8835924Z 2025-12-04T09:47:54.8836000Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8836083Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8836184Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8836375Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8836443Z res = mod(**inputs) 2025-12-04T09:47:54.8836682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8836747Z outputs = self.model( 2025-12-04T09:47:54.8836993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8837095Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8837315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8837391Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8837622Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8837711Z return func(*args, **kwargs) 2025-12-04T09:47:54.8837954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8838051Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8838294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8838360Z return func(*args, **kwargs) 2025-12-04T09:47:54.8838625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 245, in forward 2025-12-04T09:47:54.8838721Z attn_output = torch.bmm(attn_probs, value_states) 2025-12-04T09:47:54.8838724Z 2025-12-04T09:47:54.8838822Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8839019Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8839084Z res = mod(**inputs) 2025-12-04T09:47:54.8839332Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 668, in forward 2025-12-04T09:47:54.8839397Z outputs = self.model( 2025-12-04T09:47:54.8839637Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 559, in forward 2025-12-04T09:47:54.8839714Z layer_outputs = decoder_layer( 2025-12-04T09:47:54.8839936Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:47:54.8840018Z return super().__call__(*args, **kwargs) 2025-12-04T09:47:54.8840265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8840333Z return func(*args, **kwargs) 2025-12-04T09:47:54.8840636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 333, in forward 2025-12-04T09:47:54.8840732Z hidden_states, self_attn_weights = self.self_attn( 2025-12-04T09:47:54.8840964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-12-04T09:47:54.8841039Z return func(*args, **kwargs) 2025-12-04T09:47:54.8841278Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 258, in forward 2025-12-04T09:47:54.8841403Z attn_output = attn_output.reshape(bsz, tgt_len, self.embed_dim) 2025-12-04T09:47:54.8841416Z 2025-12-04T09:47:54.8841495Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8841572Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8841700Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8841789Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8841866Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8841959Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8842045Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8842128Z cudagraph partition due to non gpu ops 2025-12-04T09:47:54.8842252Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:47:54.8842475Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:47:54.8842554Z res = mod(**inputs) 2025-12-04T09:47:54.8842839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xglm/modeling_xglm.py", line 689, in forward 2025-12-04T09:47:54.8842947Z loss = self.loss_function( 2025-12-04T09:47:54.8843253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/loss/loss_utils.py", line 67, in ForCausalLMLoss 2025-12-04T09:47:54.8843444Z loss = fixed_cross_entropy(logits, shift_labels, num_items_in_batch, ignore_index, **kwargs) 2025-12-04T09:47:54.8843724Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/loss/loss_utils.py", line 36, in fixed_cross_entropy 2025-12-04T09:47:54.8843956Z loss = nn.functional.cross_entropy(source, target, ignore_index=ignore_index, reduction=reduction) 2025-12-04T09:47:54.8843961Z 2025-12-04T09:48:07.0960440Z Compilation time (from dynamo_timed): 29.810086579 2025-12-04T09:48:07.1011021Z pass 2025-12-04T09:48:07.1015112Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:48:07.1020383Z TIMING: _recursive_pre_grad_passes:0.064 _recursive_joint_graph_passes:1.08911 _recursive_post_grad_passes:0.239 async_compile.wait:0.80797 code_gen:12.00974 inductor_compile:14.9633 backend_compile:25.18203 gc:0.00018 entire_frame_compile:29.81009 total_wall_time:29.81009 2025-12-04T09:48:07.1021445Z STATS: call_* op count: 923 | FakeTensorMode.__torch_dispatch__:48065 | FakeTensor.__torch_dispatch__:7353 | ProxyTorchDispatchMode.__torch_dispatch__:9013 2025-12-04T09:48:07.1021930Z Dynamo produced 1 graphs covering 923 ops with 0 graph breaks (0 unique) 2025-12-04T09:48:10.2718339Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:48:10.2723939Z import pynvml # type: ignore[import] 2025-12-04T09:48:13.4593953Z 2025-12-04T09:48:15.8963361Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:48:15.8963833Z loading model: 0it [00:02, ?it/s] 2025-12-04T09:48:15.8964215Z cpu eval XLNetLMHeadModel 2025-12-04T09:48:18.3359151Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:48:18.8726196Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:48:19.4189768Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:48:44.6342963Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6343549Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6344014Z res = mod(**inputs) 2025-12-04T09:48:44.6353141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6353639Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6354097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1334, in forward 2025-12-04T09:48:44.6354577Z pos_emb = self.relative_positional_encoding(qlen, klen, bsz=bsz) 2025-12-04T09:48:44.6355080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1157, in relative_positional_encoding 2025-12-04T09:48:44.6355577Z pos_emb = self.positional_embedding(fwd_pos_seq, inv_freq, bsz) 2025-12-04T09:48:44.6356056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1115, in positional_embedding 2025-12-04T09:48:44.6356575Z pos_emb = torch.cat([torch.sin(sinusoid_inp), torch.cos(sinusoid_inp)], dim=-1) 2025-12-04T09:48:44.6356801Z 2025-12-04T09:48:44.6356922Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6357309Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6357646Z res = mod(**inputs) 2025-12-04T09:48:44.6358410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6358842Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6359311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1334, in forward 2025-12-04T09:48:44.6359789Z pos_emb = self.relative_positional_encoding(qlen, klen, bsz=bsz) 2025-12-04T09:48:44.6360377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1157, in relative_positional_encoding 2025-12-04T09:48:44.6360909Z pos_emb = self.positional_embedding(fwd_pos_seq, inv_freq, bsz) 2025-12-04T09:48:44.6361416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1115, in positional_embedding 2025-12-04T09:48:44.6362232Z pos_emb = torch.cat([torch.sin(sinusoid_inp), torch.cos(sinusoid_inp)], dim=-1) 2025-12-04T09:48:44.6362477Z 2025-12-04T09:48:44.6362612Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6362986Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6363314Z res = mod(**inputs) 2025-12-04T09:48:44.6363693Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6364103Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6364517Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1307, in forward 2025-12-04T09:48:44.6364930Z word_emb_k = self.word_embedding(input_ids) 2025-12-04T09:48:44.6365101Z 2025-12-04T09:48:44.6365210Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6365580Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6365908Z res = mod(**inputs) 2025-12-04T09:48:44.6366398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6366823Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6367239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6367628Z outputs = layer_module( 2025-12-04T09:48:44.6368056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6368487Z outputs = self.rel_attn( 2025-12-04T09:48:44.6368854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6369251Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6369411Z 2025-12-04T09:48:44.6369519Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6369877Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6370194Z res = mod(**inputs) 2025-12-04T09:48:44.6370555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6370955Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6371355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6371725Z outputs = layer_module( 2025-12-04T09:48:44.6372083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6372459Z outputs = self.rel_attn( 2025-12-04T09:48:44.6372985Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6373444Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6373611Z 2025-12-04T09:48:44.6373715Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6374072Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6374383Z res = mod(**inputs) 2025-12-04T09:48:44.6374743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6375159Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6375566Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6375934Z outputs = layer_module( 2025-12-04T09:48:44.6376326Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6376732Z outputs = self.rel_attn( 2025-12-04T09:48:44.6377085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6377475Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6377880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6378344Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6378534Z 2025-12-04T09:48:44.6378640Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6378999Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6379318Z res = mod(**inputs) 2025-12-04T09:48:44.6379677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6380086Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6380494Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1334, in forward 2025-12-04T09:48:44.6380938Z pos_emb = self.relative_positional_encoding(qlen, klen, bsz=bsz) 2025-12-04T09:48:44.6381433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1157, in relative_positional_encoding 2025-12-04T09:48:44.6381909Z pos_emb = self.positional_embedding(fwd_pos_seq, inv_freq, bsz) 2025-12-04T09:48:44.6382366Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1115, in positional_embedding 2025-12-04T09:48:44.6382853Z pos_emb = torch.cat([torch.sin(sinusoid_inp), torch.cos(sinusoid_inp)], dim=-1) 2025-12-04T09:48:44.6383052Z 2025-12-04T09:48:44.6383157Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6383518Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6383854Z res = mod(**inputs) 2025-12-04T09:48:44.6384206Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6384590Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6384985Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6385366Z outputs = layer_module( 2025-12-04T09:48:44.6385732Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6386107Z outputs = self.rel_attn( 2025-12-04T09:48:44.6386478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6386932Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6387146Z 2025-12-04T09:48:44.6387265Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6387628Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6387944Z res = mod(**inputs) 2025-12-04T09:48:44.6388301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6388702Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6389093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6389471Z outputs = layer_module( 2025-12-04T09:48:44.6389829Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6390213Z outputs = self.rel_attn( 2025-12-04T09:48:44.6390605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6391075Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6391264Z 2025-12-04T09:48:44.6391370Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6391736Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6392058Z res = mod(**inputs) 2025-12-04T09:48:44.6392445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6392874Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6393297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6393691Z outputs = layer_module( 2025-12-04T09:48:44.6394062Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6394461Z outputs = self.rel_attn( 2025-12-04T09:48:44.6394829Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6395222Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6395624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6396092Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6396277Z 2025-12-04T09:48:44.6396388Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6396746Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6397059Z res = mod(**inputs) 2025-12-04T09:48:44.6397423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6397833Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6398239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6398627Z outputs = layer_module( 2025-12-04T09:48:44.6399006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6399401Z outputs = self.rel_attn( 2025-12-04T09:48:44.6399783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6400214Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6400373Z 2025-12-04T09:48:44.6400487Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6400856Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6401234Z res = mod(**inputs) 2025-12-04T09:48:44.6401726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6402158Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6402594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6403041Z outputs = layer_module( 2025-12-04T09:48:44.6403438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6403863Z outputs = self.rel_attn( 2025-12-04T09:48:44.6404304Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6404716Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6405145Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6405609Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6405803Z 2025-12-04T09:48:44.6405911Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6406292Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6406646Z res = mod(**inputs) 2025-12-04T09:48:44.6407012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6407429Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6407838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6408236Z outputs = layer_module( 2025-12-04T09:48:44.6408618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6409037Z outputs = self.rel_attn( 2025-12-04T09:48:44.6409439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6409902Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6410351Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6410843Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6411020Z 2025-12-04T09:48:44.6411139Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6411512Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6411873Z res = mod(**inputs) 2025-12-04T09:48:44.6412257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6412683Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6413094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6413512Z outputs = layer_module( 2025-12-04T09:48:44.6413898Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6414297Z outputs = self.rel_attn( 2025-12-04T09:48:44.6414684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6415078Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6415489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6415922Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6416133Z 2025-12-04T09:48:44.6416232Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6416449Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6416653Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6416861Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6417107Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6417454Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6417778Z res = mod(**inputs) 2025-12-04T09:48:44.6418133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6418532Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6418921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6419315Z outputs = layer_module( 2025-12-04T09:48:44.6419702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6420097Z outputs = self.rel_attn( 2025-12-04T09:48:44.6420467Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6420894Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6421048Z 2025-12-04T09:48:44.6421161Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6421505Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6421823Z res = mod(**inputs) 2025-12-04T09:48:44.6422193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6422612Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6423002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6423383Z outputs = layer_module( 2025-12-04T09:48:44.6423751Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6424137Z outputs = self.rel_attn( 2025-12-04T09:48:44.6424497Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6424910Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6425063Z 2025-12-04T09:48:44.6425176Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6425526Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6425854Z res = mod(**inputs) 2025-12-04T09:48:44.6426209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6426598Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6426976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6427356Z outputs = layer_module( 2025-12-04T09:48:44.6427731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6428118Z outputs = self.rel_attn( 2025-12-04T09:48:44.6428500Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6428909Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6429315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6429768Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6429986Z 2025-12-04T09:48:44.6430388Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6430766Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6431089Z res = mod(**inputs) 2025-12-04T09:48:44.6431448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6431902Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6432298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6432668Z outputs = layer_module( 2025-12-04T09:48:44.6433036Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6433429Z outputs = self.rel_attn( 2025-12-04T09:48:44.6433830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6434280Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6434481Z 2025-12-04T09:48:44.6434591Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6434962Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6435294Z res = mod(**inputs) 2025-12-04T09:48:44.6435649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6436041Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6436435Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6436809Z outputs = layer_module( 2025-12-04T09:48:44.6437176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6437562Z outputs = self.rel_attn( 2025-12-04T09:48:44.6437931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6438317Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6438727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6439195Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6439388Z 2025-12-04T09:48:44.6439507Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6439884Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6440228Z res = mod(**inputs) 2025-12-04T09:48:44.6440619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6441053Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6441491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6442011Z outputs = layer_module( 2025-12-04T09:48:44.6442420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6442845Z outputs = self.rel_attn( 2025-12-04T09:48:44.6443261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6443690Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6443850Z 2025-12-04T09:48:44.6443963Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6444340Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6444748Z res = mod(**inputs) 2025-12-04T09:48:44.6445177Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6445589Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6445997Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6446404Z outputs = layer_module( 2025-12-04T09:48:44.6446782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6447174Z outputs = self.rel_attn( 2025-12-04T09:48:44.6447544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6447933Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6448352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6448809Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6448987Z 2025-12-04T09:48:44.6449093Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6449447Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6449759Z res = mod(**inputs) 2025-12-04T09:48:44.6450120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6450516Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6450905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6451283Z outputs = layer_module( 2025-12-04T09:48:44.6451650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6452031Z outputs = self.rel_attn( 2025-12-04T09:48:44.6452390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6452789Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6453209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6453654Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6453818Z 2025-12-04T09:48:44.6453923Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6454280Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6454599Z res = mod(**inputs) 2025-12-04T09:48:44.6454950Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6455359Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6455763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6456154Z outputs = layer_module( 2025-12-04T09:48:44.6456520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6456926Z outputs = self.rel_attn( 2025-12-04T09:48:44.6457292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6457699Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6458120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6458578Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6458766Z 2025-12-04T09:48:44.6458861Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6459094Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6459315Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6459531Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6459764Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6460130Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6460480Z res = mod(**inputs) 2025-12-04T09:48:44.6460851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6461253Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6461663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6462055Z outputs = layer_module( 2025-12-04T09:48:44.6462446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6462834Z outputs = self.rel_attn( 2025-12-04T09:48:44.6463212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6463629Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6463785Z 2025-12-04T09:48:44.6463894Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6464259Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6464588Z res = mod(**inputs) 2025-12-04T09:48:44.6464955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6465354Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6465755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6466148Z outputs = layer_module( 2025-12-04T09:48:44.6466523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6466912Z outputs = self.rel_attn( 2025-12-04T09:48:44.6467289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6467712Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6467869Z 2025-12-04T09:48:44.6467975Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6468343Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6468661Z res = mod(**inputs) 2025-12-04T09:48:44.6469020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6469407Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6469799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6470176Z outputs = layer_module( 2025-12-04T09:48:44.6470530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6470910Z outputs = self.rel_attn( 2025-12-04T09:48:44.6471276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6471661Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6472055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6472516Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6472714Z 2025-12-04T09:48:44.6472841Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6473188Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6473504Z res = mod(**inputs) 2025-12-04T09:48:44.6473864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6474313Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6474702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6475084Z outputs = layer_module( 2025-12-04T09:48:44.6475452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6475835Z outputs = self.rel_attn( 2025-12-04T09:48:44.6476212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6476660Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6476846Z 2025-12-04T09:48:44.6476959Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6477308Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6477636Z res = mod(**inputs) 2025-12-04T09:48:44.6478005Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6478407Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6478798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6479180Z outputs = layer_module( 2025-12-04T09:48:44.6479548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6479926Z outputs = self.rel_attn( 2025-12-04T09:48:44.6480291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6480678Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6481077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6481531Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6481824Z 2025-12-04T09:48:44.6481943Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6482332Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6482706Z res = mod(**inputs) 2025-12-04T09:48:44.6483114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6483548Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6483952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6484335Z outputs = layer_module( 2025-12-04T09:48:44.6484709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6485103Z outputs = self.rel_attn( 2025-12-04T09:48:44.6485488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6485908Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6486072Z 2025-12-04T09:48:44.6486176Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6486541Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6486886Z res = mod(**inputs) 2025-12-04T09:48:44.6487258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6487655Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6488052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6488415Z outputs = layer_module( 2025-12-04T09:48:44.6488787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6489157Z outputs = self.rel_attn( 2025-12-04T09:48:44.6489513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6489880Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6490289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6490742Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6490916Z 2025-12-04T09:48:44.6491027Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6491374Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6491691Z res = mod(**inputs) 2025-12-04T09:48:44.6492050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6492435Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6492828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6493209Z outputs = layer_module( 2025-12-04T09:48:44.6493572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6493944Z outputs = self.rel_attn( 2025-12-04T09:48:44.6494312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6494711Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6495121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6495567Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6495741Z 2025-12-04T09:48:44.6495847Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6496201Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6496514Z res = mod(**inputs) 2025-12-04T09:48:44.6496875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6497274Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6497660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6498041Z outputs = layer_module( 2025-12-04T09:48:44.6498414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6498794Z outputs = self.rel_attn( 2025-12-04T09:48:44.6499152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6499551Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6499971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6500411Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6500575Z 2025-12-04T09:48:44.6500671Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6500899Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6501109Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6501307Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6501542Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6501900Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6502245Z res = mod(**inputs) 2025-12-04T09:48:44.6502606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6502990Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6503370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6503729Z outputs = layer_module( 2025-12-04T09:48:44.6504099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6504472Z outputs = self.rel_attn( 2025-12-04T09:48:44.6504821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6505202Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6505353Z 2025-12-04T09:48:44.6505453Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6505790Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6506086Z res = mod(**inputs) 2025-12-04T09:48:44.6506430Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6506804Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6507180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6507539Z outputs = layer_module( 2025-12-04T09:48:44.6507888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6508258Z outputs = self.rel_attn( 2025-12-04T09:48:44.6508615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6509026Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6509178Z 2025-12-04T09:48:44.6509280Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6509623Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6509933Z res = mod(**inputs) 2025-12-04T09:48:44.6510285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6510676Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6511064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6511427Z outputs = layer_module( 2025-12-04T09:48:44.6511790Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6512167Z outputs = self.rel_attn( 2025-12-04T09:48:44.6512532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6512915Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6513321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6513772Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6513968Z 2025-12-04T09:48:44.6514086Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6514438Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6514749Z res = mod(**inputs) 2025-12-04T09:48:44.6515102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6515504Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6515896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6516267Z outputs = layer_module( 2025-12-04T09:48:44.6516625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6516998Z outputs = self.rel_attn( 2025-12-04T09:48:44.6517376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6517817Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6518000Z 2025-12-04T09:48:44.6518102Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6518459Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6518772Z res = mod(**inputs) 2025-12-04T09:48:44.6519123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6519514Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6519902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6520276Z outputs = layer_module( 2025-12-04T09:48:44.6520674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6521059Z outputs = self.rel_attn( 2025-12-04T09:48:44.6521429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6521912Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6522361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6522858Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6523045Z 2025-12-04T09:48:44.6523173Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6523526Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6523851Z res = mod(**inputs) 2025-12-04T09:48:44.6524218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6524614Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6524998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6525372Z outputs = layer_module( 2025-12-04T09:48:44.6525732Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6526105Z outputs = self.rel_attn( 2025-12-04T09:48:44.6526460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6526867Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6527015Z 2025-12-04T09:48:44.6527125Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6527501Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6527820Z res = mod(**inputs) 2025-12-04T09:48:44.6528211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6528605Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6528986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6529360Z outputs = layer_module( 2025-12-04T09:48:44.6529734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6530235Z outputs = self.rel_attn( 2025-12-04T09:48:44.6530621Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6531005Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6531444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6531888Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6532071Z 2025-12-04T09:48:44.6532175Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6532526Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6532836Z res = mod(**inputs) 2025-12-04T09:48:44.6533182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6533571Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6533958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6534324Z outputs = layer_module( 2025-12-04T09:48:44.6534688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6535066Z outputs = self.rel_attn( 2025-12-04T09:48:44.6535473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6535848Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6536248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6536673Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6536833Z 2025-12-04T09:48:44.6536942Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6537274Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6537580Z res = mod(**inputs) 2025-12-04T09:48:44.6537923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6538297Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6538672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6539032Z outputs = layer_module( 2025-12-04T09:48:44.6539375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6539736Z outputs = self.rel_attn( 2025-12-04T09:48:44.6540098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6540494Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6540900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6541325Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6541492Z 2025-12-04T09:48:44.6541605Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6541842Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6542042Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6542246Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6542475Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6542813Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6543151Z res = mod(**inputs) 2025-12-04T09:48:44.6543509Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6543902Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6544285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6544658Z outputs = layer_module( 2025-12-04T09:48:44.6545063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6545438Z outputs = self.rel_attn( 2025-12-04T09:48:44.6545805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6546211Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6546360Z 2025-12-04T09:48:44.6546473Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6546820Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6547136Z res = mod(**inputs) 2025-12-04T09:48:44.6547492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6547884Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6548267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6548649Z outputs = layer_module( 2025-12-04T09:48:44.6549009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6549375Z outputs = self.rel_attn( 2025-12-04T09:48:44.6549739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6550147Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6550297Z 2025-12-04T09:48:44.6550407Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6550747Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6551059Z res = mod(**inputs) 2025-12-04T09:48:44.6551408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6551793Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6552185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6552564Z outputs = layer_module( 2025-12-04T09:48:44.6552924Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6553294Z outputs = self.rel_attn( 2025-12-04T09:48:44.6553658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6554041Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6554423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6554859Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6555041Z 2025-12-04T09:48:44.6555169Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6555540Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6555847Z res = mod(**inputs) 2025-12-04T09:48:44.6556197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6556584Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6556984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6557343Z outputs = layer_module( 2025-12-04T09:48:44.6557697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6558064Z outputs = self.rel_attn( 2025-12-04T09:48:44.6558427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6558865Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6559060Z 2025-12-04T09:48:44.6559165Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6559519Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6559829Z res = mod(**inputs) 2025-12-04T09:48:44.6560189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6560584Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6560975Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6561348Z outputs = layer_module( 2025-12-04T09:48:44.6561803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6562205Z outputs = self.rel_attn( 2025-12-04T09:48:44.6562578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6562977Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6563395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6563846Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6564025Z 2025-12-04T09:48:44.6564132Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6564483Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6564800Z res = mod(**inputs) 2025-12-04T09:48:44.6565157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6565554Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6565960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6566339Z outputs = layer_module( 2025-12-04T09:48:44.6566689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6567065Z outputs = self.rel_attn( 2025-12-04T09:48:44.6567424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6567824Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6567974Z 2025-12-04T09:48:44.6568077Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6568422Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6568733Z res = mod(**inputs) 2025-12-04T09:48:44.6569118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6569505Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6569890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6570259Z outputs = layer_module( 2025-12-04T09:48:44.6570626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6571001Z outputs = self.rel_attn( 2025-12-04T09:48:44.6571361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6571737Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6572135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6572583Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6572755Z 2025-12-04T09:48:44.6572869Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6573214Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6573532Z res = mod(**inputs) 2025-12-04T09:48:44.6573887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6574286Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6574665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6575042Z outputs = layer_module( 2025-12-04T09:48:44.6575403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6575774Z outputs = self.rel_attn( 2025-12-04T09:48:44.6576139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6576538Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6576955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6577390Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6577564Z 2025-12-04T09:48:44.6577670Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6578025Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6578348Z res = mod(**inputs) 2025-12-04T09:48:44.6578698Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6579092Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6579486Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6579856Z outputs = layer_module( 2025-12-04T09:48:44.6580216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6580595Z outputs = self.rel_attn( 2025-12-04T09:48:44.6580963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6581353Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6581776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6582216Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6582379Z 2025-12-04T09:48:44.6582486Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6582704Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6582912Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6583117Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6583339Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6583689Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6584021Z res = mod(**inputs) 2025-12-04T09:48:44.6584368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6584756Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6585141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6585514Z outputs = layer_module( 2025-12-04T09:48:44.6585879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6586258Z outputs = self.rel_attn( 2025-12-04T09:48:44.6586621Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6587025Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6587172Z 2025-12-04T09:48:44.6587274Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6587625Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6587939Z res = mod(**inputs) 2025-12-04T09:48:44.6588284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6588672Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6589061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6589439Z outputs = layer_module( 2025-12-04T09:48:44.6589792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6590167Z outputs = self.rel_attn( 2025-12-04T09:48:44.6590528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6590930Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6591089Z 2025-12-04T09:48:44.6591188Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6591537Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6591845Z res = mod(**inputs) 2025-12-04T09:48:44.6592192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6592584Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6592975Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6593347Z outputs = layer_module( 2025-12-04T09:48:44.6593700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6594080Z outputs = self.rel_attn( 2025-12-04T09:48:44.6594442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6594814Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6595210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6595658Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6595836Z 2025-12-04T09:48:44.6595974Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6596322Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6596628Z res = mod(**inputs) 2025-12-04T09:48:44.6596973Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6597343Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6597738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6598098Z outputs = layer_module( 2025-12-04T09:48:44.6598451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6598813Z outputs = self.rel_attn( 2025-12-04T09:48:44.6599187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6599625Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6599804Z 2025-12-04T09:48:44.6599913Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6600255Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6600579Z res = mod(**inputs) 2025-12-04T09:48:44.6600926Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6601294Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6601759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6602141Z outputs = layer_module( 2025-12-04T09:48:44.6602506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6602886Z outputs = self.rel_attn( 2025-12-04T09:48:44.6603266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6603672Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6604060Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6604523Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6604712Z 2025-12-04T09:48:44.6604818Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6605187Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6605493Z res = mod(**inputs) 2025-12-04T09:48:44.6605844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6606227Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6606609Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6606970Z outputs = layer_module( 2025-12-04T09:48:44.6607323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6607693Z outputs = self.rel_attn( 2025-12-04T09:48:44.6608044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6608441Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6608595Z 2025-12-04T09:48:44.6608696Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6609042Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6609346Z res = mod(**inputs) 2025-12-04T09:48:44.6609755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6610144Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6610529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6610891Z outputs = layer_module( 2025-12-04T09:48:44.6611265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6611639Z outputs = self.rel_attn( 2025-12-04T09:48:44.6611992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6612370Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6612776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6613221Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6613390Z 2025-12-04T09:48:44.6613492Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6613838Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6614146Z res = mod(**inputs) 2025-12-04T09:48:44.6614490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6614876Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6615257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6615627Z outputs = layer_module( 2025-12-04T09:48:44.6615972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6616342Z outputs = self.rel_attn( 2025-12-04T09:48:44.6616730Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6617121Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6617525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6617956Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6618116Z 2025-12-04T09:48:44.6618225Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6618562Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6618871Z res = mod(**inputs) 2025-12-04T09:48:44.6619217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6619605Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6619978Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6620347Z outputs = layer_module( 2025-12-04T09:48:44.6620699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6621066Z outputs = self.rel_attn( 2025-12-04T09:48:44.6621411Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6621796Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6622199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6622621Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6622787Z 2025-12-04T09:48:44.6622867Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6624055Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6624264Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6624460Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6624690Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6625043Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6625369Z res = mod(**inputs) 2025-12-04T09:48:44.6625727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6626122Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6626515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6626883Z outputs = layer_module( 2025-12-04T09:48:44.6627261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6627633Z outputs = self.rel_attn( 2025-12-04T09:48:44.6627985Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6628385Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6628540Z 2025-12-04T09:48:44.6628643Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6629004Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6629303Z res = mod(**inputs) 2025-12-04T09:48:44.6629645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6630020Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6630557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6630928Z outputs = layer_module( 2025-12-04T09:48:44.6631280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6631649Z outputs = self.rel_attn( 2025-12-04T09:48:44.6632002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6632413Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6632562Z 2025-12-04T09:48:44.6632675Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6633030Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6633343Z res = mod(**inputs) 2025-12-04T09:48:44.6633706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6634105Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6634487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6634861Z outputs = layer_module( 2025-12-04T09:48:44.6635217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6635587Z outputs = self.rel_attn( 2025-12-04T09:48:44.6635944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6636335Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6636733Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6637188Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6637375Z 2025-12-04T09:48:44.6637527Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6637902Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6638215Z res = mod(**inputs) 2025-12-04T09:48:44.6638558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6638949Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6639357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6639724Z outputs = layer_module( 2025-12-04T09:48:44.6640071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6640443Z outputs = self.rel_attn( 2025-12-04T09:48:44.6640832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6641258Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6641446Z 2025-12-04T09:48:44.6641548Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6641977Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6642324Z res = mod(**inputs) 2025-12-04T09:48:44.6642705Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6643138Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6643564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6643965Z outputs = layer_module( 2025-12-04T09:48:44.6644330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6644745Z outputs = self.rel_attn( 2025-12-04T09:48:44.6645147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6645559Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6646008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6646509Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6646705Z 2025-12-04T09:48:44.6646826Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6647206Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6647555Z res = mod(**inputs) 2025-12-04T09:48:44.6647944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6648370Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6648800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6649208Z outputs = layer_module( 2025-12-04T09:48:44.6649604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6650021Z outputs = self.rel_attn( 2025-12-04T09:48:44.6650433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6650898Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6651052Z 2025-12-04T09:48:44.6651159Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6651496Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6651805Z res = mod(**inputs) 2025-12-04T09:48:44.6652194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6652577Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6652964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6653335Z outputs = layer_module( 2025-12-04T09:48:44.6653706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6654069Z outputs = self.rel_attn( 2025-12-04T09:48:44.6654426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6654802Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6655182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6655651Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6655831Z 2025-12-04T09:48:44.6655934Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6656283Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6656585Z res = mod(**inputs) 2025-12-04T09:48:44.6656934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6657320Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6657701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6658062Z outputs = layer_module( 2025-12-04T09:48:44.6658417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6658792Z outputs = self.rel_attn( 2025-12-04T09:48:44.6659144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6659541Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6659947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6660378Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6660536Z 2025-12-04T09:48:44.6660637Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6660988Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6661297Z res = mod(**inputs) 2025-12-04T09:48:44.6661647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6662030Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6662412Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6662780Z outputs = layer_module( 2025-12-04T09:48:44.6663128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6663508Z outputs = self.rel_attn( 2025-12-04T09:48:44.6663856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6664235Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6664625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6665045Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6665201Z 2025-12-04T09:48:44.6665288Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6665520Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6665727Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6665932Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6666160Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6666500Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6666834Z res = mod(**inputs) 2025-12-04T09:48:44.6667190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6667572Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6667960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6668330Z outputs = layer_module( 2025-12-04T09:48:44.6668704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6669072Z outputs = self.rel_attn( 2025-12-04T09:48:44.6669432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6669836Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6669983Z 2025-12-04T09:48:44.6670085Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6670434Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6670749Z res = mod(**inputs) 2025-12-04T09:48:44.6671099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6671485Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6671874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6672249Z outputs = layer_module( 2025-12-04T09:48:44.6672607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6672975Z outputs = self.rel_attn( 2025-12-04T09:48:44.6673340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6673744Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6673895Z 2025-12-04T09:48:44.6673996Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6674347Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6674659Z res = mod(**inputs) 2025-12-04T09:48:44.6675008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6675390Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6675779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6676154Z outputs = layer_module( 2025-12-04T09:48:44.6676492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6676854Z outputs = self.rel_attn( 2025-12-04T09:48:44.6677207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6677575Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6677949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6678389Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6678568Z 2025-12-04T09:48:44.6678682Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6679060Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6679366Z res = mod(**inputs) 2025-12-04T09:48:44.6679718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6680104Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6680494Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6680861Z outputs = layer_module( 2025-12-04T09:48:44.6681216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6681590Z outputs = self.rel_attn( 2025-12-04T09:48:44.6682064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6682589Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6682795Z 2025-12-04T09:48:44.6682920Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6683327Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6683650Z res = mod(**inputs) 2025-12-04T09:48:44.6684015Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6684408Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6684794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6685173Z outputs = layer_module( 2025-12-04T09:48:44.6685534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6685909Z outputs = self.rel_attn( 2025-12-04T09:48:44.6686314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6686687Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6687075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6687515Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6687699Z 2025-12-04T09:48:44.6687802Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6688154Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6688470Z res = mod(**inputs) 2025-12-04T09:48:44.6688813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6689203Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6689592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6689963Z outputs = layer_module( 2025-12-04T09:48:44.6690312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6690687Z outputs = self.rel_attn( 2025-12-04T09:48:44.6691050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6691443Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6691598Z 2025-12-04T09:48:44.6691701Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6692047Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6692357Z res = mod(**inputs) 2025-12-04T09:48:44.6692756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6693142Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6693521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6693884Z outputs = layer_module( 2025-12-04T09:48:44.6694254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6694626Z outputs = self.rel_attn( 2025-12-04T09:48:44.6694987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6695404Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6695783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6696227Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6696395Z 2025-12-04T09:48:44.6696503Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6696841Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6697150Z res = mod(**inputs) 2025-12-04T09:48:44.6697498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6697874Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6698252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6698616Z outputs = layer_module( 2025-12-04T09:48:44.6698967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6699328Z outputs = self.rel_attn( 2025-12-04T09:48:44.6699688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6700075Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6700472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6700900Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6701067Z 2025-12-04T09:48:44.6701166Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6701509Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6701807Z res = mod(**inputs) 2025-12-04T09:48:44.6702154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6702533Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6702916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6703274Z outputs = layer_module( 2025-12-04T09:48:44.6703626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6703994Z outputs = self.rel_attn( 2025-12-04T09:48:44.6704340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6704724Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6705126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6705549Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6705706Z 2025-12-04T09:48:44.6705783Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6706008Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6706225Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6706418Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6706947Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6707296Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6707607Z res = mod(**inputs) 2025-12-04T09:48:44.6707983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6708382Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6708759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6709122Z outputs = layer_module( 2025-12-04T09:48:44.6709480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6709842Z outputs = self.rel_attn( 2025-12-04T09:48:44.6710193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6710571Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6710722Z 2025-12-04T09:48:44.6710820Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6711162Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6711470Z res = mod(**inputs) 2025-12-04T09:48:44.6711812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6712197Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6712580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6712947Z outputs = layer_module( 2025-12-04T09:48:44.6713301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6713670Z outputs = self.rel_attn( 2025-12-04T09:48:44.6714029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6714428Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6714582Z 2025-12-04T09:48:44.6714683Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6715032Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6715347Z res = mod(**inputs) 2025-12-04T09:48:44.6715691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6716079Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6716469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6716838Z outputs = layer_module( 2025-12-04T09:48:44.6717196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6717568Z outputs = self.rel_attn( 2025-12-04T09:48:44.6717927Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6718295Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6718685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6719135Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6719315Z 2025-12-04T09:48:44.6719426Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6719803Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6720128Z res = mod(**inputs) 2025-12-04T09:48:44.6720481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6720860Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6721264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6721723Z outputs = layer_module( 2025-12-04T09:48:44.6722103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6722472Z outputs = self.rel_attn( 2025-12-04T09:48:44.6722836Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6723299Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6723483Z 2025-12-04T09:48:44.6723587Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6723941Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6724254Z res = mod(**inputs) 2025-12-04T09:48:44.6724604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6724985Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6725372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6725746Z outputs = layer_module( 2025-12-04T09:48:44.6726100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6726465Z outputs = self.rel_attn( 2025-12-04T09:48:44.6726825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6727204Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6727584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6728027Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6728209Z 2025-12-04T09:48:44.6728321Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6728660Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6728957Z res = mod(**inputs) 2025-12-04T09:48:44.6729296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6729678Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6730044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6730566Z outputs = layer_module( 2025-12-04T09:48:44.6730919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6731288Z outputs = self.rel_attn( 2025-12-04T09:48:44.6731641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6732037Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6732041Z 2025-12-04T09:48:44.6732150Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6732338Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6732403Z res = mod(**inputs) 2025-12-04T09:48:44.6732706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6732807Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6733046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6733120Z outputs = layer_module( 2025-12-04T09:48:44.6733357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6733452Z outputs = self.rel_attn( 2025-12-04T09:48:44.6733687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6733758Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6734023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6734168Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6734172Z 2025-12-04T09:48:44.6734274Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6734469Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6734533Z res = mod(**inputs) 2025-12-04T09:48:44.6734779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6734860Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6735101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6735177Z outputs = layer_module( 2025-12-04T09:48:44.6735413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6735490Z outputs = self.rel_attn( 2025-12-04T09:48:44.6735728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6735816Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6736079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6736185Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6736190Z 2025-12-04T09:48:44.6736287Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6736479Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6736541Z res = mod(**inputs) 2025-12-04T09:48:44.6736786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6736865Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6737103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6737178Z outputs = layer_module( 2025-12-04T09:48:44.6737414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6737485Z outputs = self.rel_attn( 2025-12-04T09:48:44.6737722Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6737807Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6738077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6738181Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6738185Z 2025-12-04T09:48:44.6738260Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6738365Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6738455Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6738537Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6738634Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6738822Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6738889Z res = mod(**inputs) 2025-12-04T09:48:44.6739146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6739226Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6739474Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6739538Z outputs = layer_module( 2025-12-04T09:48:44.6739803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6739873Z outputs = self.rel_attn( 2025-12-04T09:48:44.6740109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6740208Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6740212Z 2025-12-04T09:48:44.6740307Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6740492Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6740562Z res = mod(**inputs) 2025-12-04T09:48:44.6740810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6740897Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6741142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6741210Z outputs = layer_module( 2025-12-04T09:48:44.6741462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6741527Z outputs = self.rel_attn( 2025-12-04T09:48:44.6741778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6741876Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6741879Z 2025-12-04T09:48:44.6741977Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6742176Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6742237Z res = mod(**inputs) 2025-12-04T09:48:44.6742492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6742577Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6742818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6742889Z outputs = layer_module( 2025-12-04T09:48:44.6743128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6743193Z outputs = self.rel_attn( 2025-12-04T09:48:44.6743441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6743510Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6743765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6743899Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6743902Z 2025-12-04T09:48:44.6744001Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6744227Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6744292Z res = mod(**inputs) 2025-12-04T09:48:44.6744532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6744618Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6744874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6744947Z outputs = layer_module( 2025-12-04T09:48:44.6745184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6745248Z outputs = self.rel_attn( 2025-12-04T09:48:44.6745490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6745648Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6745651Z 2025-12-04T09:48:44.6745758Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6745940Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6746001Z res = mod(**inputs) 2025-12-04T09:48:44.6746250Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6746329Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6746569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6746641Z outputs = layer_module( 2025-12-04T09:48:44.6746884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6746958Z outputs = self.rel_attn( 2025-12-04T09:48:44.6747201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6747270Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6747532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6747654Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6747658Z 2025-12-04T09:48:44.6747764Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6747955Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6748017Z res = mod(**inputs) 2025-12-04T09:48:44.6748266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6748344Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6748585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6748659Z outputs = layer_module( 2025-12-04T09:48:44.6748894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6748965Z outputs = self.rel_attn( 2025-12-04T09:48:44.6749213Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6749307Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6749310Z 2025-12-04T09:48:44.6749412Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6749595Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6749655Z res = mod(**inputs) 2025-12-04T09:48:44.6749923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6750020Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6750264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6750328Z outputs = layer_module( 2025-12-04T09:48:44.6750565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6750653Z outputs = self.rel_attn( 2025-12-04T09:48:44.6750890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6750966Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6751218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6751350Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6751355Z 2025-12-04T09:48:44.6751460Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6751647Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6751707Z res = mod(**inputs) 2025-12-04T09:48:44.6751954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6752034Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6752278Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6752341Z outputs = layer_module( 2025-12-04T09:48:44.6752577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6752649Z outputs = self.rel_attn( 2025-12-04T09:48:44.6752889Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6752981Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6753235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6753341Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6753345Z 2025-12-04T09:48:44.6753449Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6753631Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6753692Z res = mod(**inputs) 2025-12-04T09:48:44.6753939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6754018Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6754270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6754336Z outputs = layer_module( 2025-12-04T09:48:44.6754578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6754650Z outputs = self.rel_attn( 2025-12-04T09:48:44.6754894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6754980Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6755278Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6755384Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6755388Z 2025-12-04T09:48:44.6755471Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6755568Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6755659Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6755741Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6755839Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6756035Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6756098Z res = mod(**inputs) 2025-12-04T09:48:44.6756364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6756452Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6756701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6756768Z outputs = layer_module( 2025-12-04T09:48:44.6757037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6757109Z outputs = self.rel_attn( 2025-12-04T09:48:44.6757367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6757462Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6757465Z 2025-12-04T09:48:44.6757564Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6757766Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6757830Z res = mod(**inputs) 2025-12-04T09:48:44.6758081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6758167Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6758411Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6758487Z outputs = layer_module( 2025-12-04T09:48:44.6758731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6758797Z outputs = self.rel_attn( 2025-12-04T09:48:44.6759051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6759149Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6759153Z 2025-12-04T09:48:44.6759260Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6759450Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6759512Z res = mod(**inputs) 2025-12-04T09:48:44.6759767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6759847Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6760095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6760171Z outputs = layer_module( 2025-12-04T09:48:44.6760417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6760491Z outputs = self.rel_attn( 2025-12-04T09:48:44.6760741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6760812Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6761087Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6761219Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6761222Z 2025-12-04T09:48:44.6761336Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6761564Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6761719Z res = mod(**inputs) 2025-12-04T09:48:44.6762008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6762099Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6762386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6762494Z outputs = layer_module( 2025-12-04T09:48:44.6762776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6762860Z outputs = self.rel_attn( 2025-12-04T09:48:44.6763146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6763312Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6763317Z 2025-12-04T09:48:44.6763440Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6763662Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6763735Z res = mod(**inputs) 2025-12-04T09:48:44.6763990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6764075Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6764340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6764409Z outputs = layer_module( 2025-12-04T09:48:44.6764659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6764736Z outputs = self.rel_attn( 2025-12-04T09:48:44.6764992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6765073Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6765343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6765472Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6765477Z 2025-12-04T09:48:44.6765586Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6765782Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6765846Z res = mod(**inputs) 2025-12-04T09:48:44.6766165Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6766250Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6766525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6766593Z outputs = layer_module( 2025-12-04T09:48:44.6766850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6766932Z outputs = self.rel_attn( 2025-12-04T09:48:44.6767190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6767300Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6767303Z 2025-12-04T09:48:44.6767407Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6767605Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6767678Z res = mod(**inputs) 2025-12-04T09:48:44.6767988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6768087Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6768345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6768412Z outputs = layer_module( 2025-12-04T09:48:44.6768674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6768758Z outputs = self.rel_attn( 2025-12-04T09:48:44.6769008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6769086Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6769357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6769496Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6769502Z 2025-12-04T09:48:44.6769602Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6769790Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6769859Z res = mod(**inputs) 2025-12-04T09:48:44.6770105Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6770187Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6770443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6770510Z outputs = layer_module( 2025-12-04T09:48:44.6770764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6770832Z outputs = self.rel_attn( 2025-12-04T09:48:44.6771094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6771188Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6771452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6771567Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6771572Z 2025-12-04T09:48:44.6771674Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6771870Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6771939Z res = mod(**inputs) 2025-12-04T09:48:44.6772231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6772313Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6772576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6772643Z outputs = layer_module( 2025-12-04T09:48:44.6772899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6772967Z outputs = self.rel_attn( 2025-12-04T09:48:44.6773216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6773314Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6773585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6773696Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6773706Z 2025-12-04T09:48:44.6773785Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6773881Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6773983Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6774063Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6774166Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6774367Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6774432Z res = mod(**inputs) 2025-12-04T09:48:44.6774701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6774790Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6775042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6775118Z outputs = layer_module( 2025-12-04T09:48:44.6775386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6775459Z outputs = self.rel_attn( 2025-12-04T09:48:44.6775719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6775818Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6775822Z 2025-12-04T09:48:44.6775930Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6776127Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6776192Z res = mod(**inputs) 2025-12-04T09:48:44.6776454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6776537Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6776788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6776867Z outputs = layer_module( 2025-12-04T09:48:44.6777125Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6777201Z outputs = self.rel_attn( 2025-12-04T09:48:44.6777456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6777554Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6777558Z 2025-12-04T09:48:44.6777667Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6777860Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6777931Z res = mod(**inputs) 2025-12-04T09:48:44.6778183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6778265Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6778525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6778591Z outputs = layer_module( 2025-12-04T09:48:44.6778844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6778920Z outputs = self.rel_attn( 2025-12-04T09:48:44.6779170Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6779250Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6779514Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6779645Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6779648Z 2025-12-04T09:48:44.6779756Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6779985Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6780059Z res = mod(**inputs) 2025-12-04T09:48:44.6780313Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6780394Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6780656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6780741Z outputs = layer_module( 2025-12-04T09:48:44.6780991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6781068Z outputs = self.rel_attn( 2025-12-04T09:48:44.6781326Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6781482Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6781486Z 2025-12-04T09:48:44.6781590Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6781787Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6781859Z res = mod(**inputs) 2025-12-04T09:48:44.6782110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6782192Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6782450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6782517Z outputs = layer_module( 2025-12-04T09:48:44.6782821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6782889Z outputs = self.rel_attn( 2025-12-04T09:48:44.6783140Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6783219Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6783490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6783622Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6783627Z 2025-12-04T09:48:44.6783727Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6783917Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6783988Z res = mod(**inputs) 2025-12-04T09:48:44.6784242Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6784320Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6784572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6784639Z outputs = layer_module( 2025-12-04T09:48:44.6784889Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6784955Z outputs = self.rel_attn( 2025-12-04T09:48:44.6785195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6785300Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6785304Z 2025-12-04T09:48:44.6785402Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6785598Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6785660Z res = mod(**inputs) 2025-12-04T09:48:44.6785921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6786028Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6786272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6786338Z outputs = layer_module( 2025-12-04T09:48:44.6786589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6786671Z outputs = self.rel_attn( 2025-12-04T09:48:44.6786921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6786991Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6787249Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6787402Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6787407Z 2025-12-04T09:48:44.6787511Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6787707Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6787769Z res = mod(**inputs) 2025-12-04T09:48:44.6788012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6788101Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6788341Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6788405Z outputs = layer_module( 2025-12-04T09:48:44.6788654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6788720Z outputs = self.rel_attn( 2025-12-04T09:48:44.6788972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6789059Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6789319Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6789433Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6789438Z 2025-12-04T09:48:44.6789536Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6789731Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6789794Z res = mod(**inputs) 2025-12-04T09:48:44.6790037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6790126Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6790372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6790439Z outputs = layer_module( 2025-12-04T09:48:44.6790686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6790753Z outputs = self.rel_attn( 2025-12-04T09:48:44.6791000Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6791085Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6791346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6791463Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6791467Z 2025-12-04T09:48:44.6791546Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6791624Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6791739Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6791815Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6791920Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6792108Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6792170Z res = mod(**inputs) 2025-12-04T09:48:44.6792441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6792522Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6792765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6792838Z outputs = layer_module( 2025-12-04T09:48:44.6793080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6793172Z outputs = self.rel_attn( 2025-12-04T09:48:44.6793414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6793510Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6793513Z 2025-12-04T09:48:44.6793620Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6793810Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6793882Z res = mod(**inputs) 2025-12-04T09:48:44.6794127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6794206Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6794456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6794522Z outputs = layer_module( 2025-12-04T09:48:44.6794766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6794840Z outputs = self.rel_attn( 2025-12-04T09:48:44.6795084Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6795188Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6795192Z 2025-12-04T09:48:44.6795291Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6795484Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6795556Z res = mod(**inputs) 2025-12-04T09:48:44.6795801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6795887Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6796135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6796200Z outputs = layer_module( 2025-12-04T09:48:44.6796452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6796518Z outputs = self.rel_attn( 2025-12-04T09:48:44.6796761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6796841Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6797097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6797230Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6797233Z 2025-12-04T09:48:44.6797330Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6797554Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6797625Z res = mod(**inputs) 2025-12-04T09:48:44.6797871Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6797950Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6798202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6798285Z outputs = layer_module( 2025-12-04T09:48:44.6798544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6798610Z outputs = self.rel_attn( 2025-12-04T09:48:44.6798860Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6799006Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6799012Z 2025-12-04T09:48:44.6799112Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6799306Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6799368Z res = mod(**inputs) 2025-12-04T09:48:44.6799614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6799702Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6799945Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6800010Z outputs = layer_module( 2025-12-04T09:48:44.6800259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6800327Z outputs = self.rel_attn( 2025-12-04T09:48:44.6800582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6800654Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6800918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6801053Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6801057Z 2025-12-04T09:48:44.6801158Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6801360Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6801424Z res = mod(**inputs) 2025-12-04T09:48:44.6801763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6801860Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6802122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6802192Z outputs = layer_module( 2025-12-04T09:48:44.6802458Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6802530Z outputs = self.rel_attn( 2025-12-04T09:48:44.6802798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6802901Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6802905Z 2025-12-04T09:48:44.6803014Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6803227Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6803295Z res = mod(**inputs) 2025-12-04T09:48:44.6803567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6803692Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6803948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6804026Z outputs = layer_module( 2025-12-04T09:48:44.6804284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6804381Z outputs = self.rel_attn( 2025-12-04T09:48:44.6804643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6804716Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6804990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6805130Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6805135Z 2025-12-04T09:48:44.6805238Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6805444Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6805509Z res = mod(**inputs) 2025-12-04T09:48:44.6805769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6805853Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6806106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6806184Z outputs = layer_module( 2025-12-04T09:48:44.6806432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6806500Z outputs = self.rel_attn( 2025-12-04T09:48:44.6806758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6806848Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6807125Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6807235Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6807240Z 2025-12-04T09:48:44.6807343Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6807548Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6807614Z res = mod(**inputs) 2025-12-04T09:48:44.6807866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6807957Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6808211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6808286Z outputs = layer_module( 2025-12-04T09:48:44.6808535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6808604Z outputs = self.rel_attn( 2025-12-04T09:48:44.6808858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6808950Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6809227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6809338Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6809341Z 2025-12-04T09:48:44.6809420Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6809509Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6809639Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6809717Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6809825Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6810019Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6810091Z res = mod(**inputs) 2025-12-04T09:48:44.6810346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6810454Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6810713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6810780Z outputs = layer_module( 2025-12-04T09:48:44.6811029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6811121Z outputs = self.rel_attn( 2025-12-04T09:48:44.6811370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6811475Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6811478Z 2025-12-04T09:48:44.6811581Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6811773Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6811846Z res = mod(**inputs) 2025-12-04T09:48:44.6812097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6812188Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6812440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6812506Z outputs = layer_module( 2025-12-04T09:48:44.6812764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6812832Z outputs = self.rel_attn( 2025-12-04T09:48:44.6813079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6813186Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6813190Z 2025-12-04T09:48:44.6813292Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6813491Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6813555Z res = mod(**inputs) 2025-12-04T09:48:44.6813806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6813896Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6814150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6814220Z outputs = layer_module( 2025-12-04T09:48:44.6814475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6814544Z outputs = self.rel_attn( 2025-12-04T09:48:44.6814798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6814874Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6815140Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6815277Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6815280Z 2025-12-04T09:48:44.6815381Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6815615Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6815680Z res = mod(**inputs) 2025-12-04T09:48:44.6815933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6816021Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6816273Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6816355Z outputs = layer_module( 2025-12-04T09:48:44.6816621Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6816688Z outputs = self.rel_attn( 2025-12-04T09:48:44.6816945Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6817088Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6817092Z 2025-12-04T09:48:44.6817193Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6817392Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6817455Z res = mod(**inputs) 2025-12-04T09:48:44.6817711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6817792Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6818037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6818110Z outputs = layer_module( 2025-12-04T09:48:44.6818354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6818420Z outputs = self.rel_attn( 2025-12-04T09:48:44.6818677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6818746Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6819017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6819141Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6819145Z 2025-12-04T09:48:44.6819246Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6819447Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6819509Z res = mod(**inputs) 2025-12-04T09:48:44.6819764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6819845Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6820093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6820166Z outputs = layer_module( 2025-12-04T09:48:44.6820413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6820479Z outputs = self.rel_attn( 2025-12-04T09:48:44.6820733Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6820831Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6820834Z 2025-12-04T09:48:44.6820941Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6821131Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6821194Z res = mod(**inputs) 2025-12-04T09:48:44.6821447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6821562Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6821815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6821880Z outputs = layer_module( 2025-12-04T09:48:44.6822122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6822210Z outputs = self.rel_attn( 2025-12-04T09:48:44.6822455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6822526Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6822793Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6822927Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6822933Z 2025-12-04T09:48:44.6823041Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6823232Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6823294Z res = mod(**inputs) 2025-12-04T09:48:44.6823546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6823626Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6823878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6823943Z outputs = layer_module( 2025-12-04T09:48:44.6824187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6824262Z outputs = self.rel_attn( 2025-12-04T09:48:44.6824508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6824595Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6824869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6824978Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6824983Z 2025-12-04T09:48:44.6825089Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6825280Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6825344Z res = mod(**inputs) 2025-12-04T09:48:44.6825595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6825676Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6825921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6825995Z outputs = layer_module( 2025-12-04T09:48:44.6826236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6826310Z outputs = self.rel_attn( 2025-12-04T09:48:44.6826555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6826642Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6826913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6827021Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6827025Z 2025-12-04T09:48:44.6827108Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6827187Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6827283Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6827384Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6827526Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6827750Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6827826Z res = mod(**inputs) 2025-12-04T09:48:44.6828071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6828182Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6828437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6828503Z outputs = layer_module( 2025-12-04T09:48:44.6828750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6828832Z outputs = self.rel_attn( 2025-12-04T09:48:44.6829074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6829178Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6829181Z 2025-12-04T09:48:44.6829281Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6829476Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6829541Z res = mod(**inputs) 2025-12-04T09:48:44.6829788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6829876Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6830247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6830330Z outputs = layer_module( 2025-12-04T09:48:44.6830579Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6830647Z outputs = self.rel_attn( 2025-12-04T09:48:44.6830899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6830995Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6831001Z 2025-12-04T09:48:44.6831100Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6831299Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6831362Z res = mod(**inputs) 2025-12-04T09:48:44.6831615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6831695Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6831941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6832017Z outputs = layer_module( 2025-12-04T09:48:44.6832262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6832329Z outputs = self.rel_attn( 2025-12-04T09:48:44.6832581Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6832653Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6832923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6833050Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6833054Z 2025-12-04T09:48:44.6833153Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6833391Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6833484Z res = mod(**inputs) 2025-12-04T09:48:44.6833734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6833814Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6834056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6834153Z outputs = layer_module( 2025-12-04T09:48:44.6834399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6834466Z outputs = self.rel_attn( 2025-12-04T09:48:44.6834719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6834866Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6834870Z 2025-12-04T09:48:44.6834983Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6835175Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6835246Z res = mod(**inputs) 2025-12-04T09:48:44.6835502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6835583Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6835835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6835902Z outputs = layer_module( 2025-12-04T09:48:44.6836142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6836216Z outputs = self.rel_attn( 2025-12-04T09:48:44.6836463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6836534Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6836798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6836922Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6836927Z 2025-12-04T09:48:44.6837032Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6837224Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6837285Z res = mod(**inputs) 2025-12-04T09:48:44.6837539Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6837619Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6837876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6837942Z outputs = layer_module( 2025-12-04T09:48:44.6838186Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6838263Z outputs = self.rel_attn( 2025-12-04T09:48:44.6838506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6838603Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6838613Z 2025-12-04T09:48:44.6838712Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6838904Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6838975Z res = mod(**inputs) 2025-12-04T09:48:44.6839220Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6839333Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6839585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6839651Z outputs = layer_module( 2025-12-04T09:48:44.6839902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6839984Z outputs = self.rel_attn( 2025-12-04T09:48:44.6840228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6840306Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6840564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6840706Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6840711Z 2025-12-04T09:48:44.6840823Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6841015Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6841086Z res = mod(**inputs) 2025-12-04T09:48:44.6841338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6841420Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6841746Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6841820Z outputs = layer_module( 2025-12-04T09:48:44.6842073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6842149Z outputs = self.rel_attn( 2025-12-04T09:48:44.6842444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6842550Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6842857Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6842982Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6842989Z 2025-12-04T09:48:44.6843111Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6843334Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6843416Z res = mod(**inputs) 2025-12-04T09:48:44.6843704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6843799Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6844077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6844146Z outputs = layer_module( 2025-12-04T09:48:44.6844410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6844488Z outputs = self.rel_attn( 2025-12-04T09:48:44.6844749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6844847Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6845123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6845234Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6845237Z 2025-12-04T09:48:44.6845327Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6845407Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6845522Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6845616Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6845720Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6845925Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6845989Z res = mod(**inputs) 2025-12-04T09:48:44.6846244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6846352Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6846605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6846681Z outputs = layer_module( 2025-12-04T09:48:44.6846932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6847018Z outputs = self.rel_attn( 2025-12-04T09:48:44.6847279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6847375Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6847379Z 2025-12-04T09:48:44.6847480Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6847682Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6847748Z res = mod(**inputs) 2025-12-04T09:48:44.6848010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6848092Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6848343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6848416Z outputs = layer_module( 2025-12-04T09:48:44.6848671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6848739Z outputs = self.rel_attn( 2025-12-04T09:48:44.6848996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6849094Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6849099Z 2025-12-04T09:48:44.6849208Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6849403Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6849467Z res = mod(**inputs) 2025-12-04T09:48:44.6849730Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6849812Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6850077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6850146Z outputs = layer_module( 2025-12-04T09:48:44.6850397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6850474Z outputs = self.rel_attn( 2025-12-04T09:48:44.6850724Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6850795Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6851066Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6851195Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6851199Z 2025-12-04T09:48:44.6851309Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6851504Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6851602Z res = mod(**inputs) 2025-12-04T09:48:44.6851866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6851949Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6852207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6852289Z outputs = layer_module( 2025-12-04T09:48:44.6852539Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6852616Z outputs = self.rel_attn( 2025-12-04T09:48:44.6852866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6853012Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6853024Z 2025-12-04T09:48:44.6853130Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6853327Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6853395Z res = mod(**inputs) 2025-12-04T09:48:44.6853647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6853730Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6853989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6854057Z outputs = layer_module( 2025-12-04T09:48:44.6854315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6854383Z outputs = self.rel_attn( 2025-12-04T09:48:44.6854638Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6854718Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6854985Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6855112Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6855124Z 2025-12-04T09:48:44.6855226Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6855422Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6855494Z res = mod(**inputs) 2025-12-04T09:48:44.6855748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6855829Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6856095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6856163Z outputs = layer_module( 2025-12-04T09:48:44.6856413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6856481Z outputs = self.rel_attn( 2025-12-04T09:48:44.6856725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6856829Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6856833Z 2025-12-04T09:48:44.6856933Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6857120Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6857189Z res = mod(**inputs) 2025-12-04T09:48:44.6857433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6857561Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6857807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6857873Z outputs = layer_module( 2025-12-04T09:48:44.6858122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6858206Z outputs = self.rel_attn( 2025-12-04T09:48:44.6858454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6858531Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6858787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6858913Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6858934Z 2025-12-04T09:48:44.6859036Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6859228Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6859302Z res = mod(**inputs) 2025-12-04T09:48:44.6859550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6859641Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6859886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6859954Z outputs = layer_module( 2025-12-04T09:48:44.6860205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6860276Z outputs = self.rel_attn( 2025-12-04T09:48:44.6860523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6860622Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6860887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6861006Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6861011Z 2025-12-04T09:48:44.6861114Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6861307Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6861381Z res = mod(**inputs) 2025-12-04T09:48:44.6861630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6861721Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6861971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6862044Z outputs = layer_module( 2025-12-04T09:48:44.6862299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6862370Z outputs = self.rel_attn( 2025-12-04T09:48:44.6862618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6862716Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6862981Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6863099Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6863102Z 2025-12-04T09:48:44.6863186Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6863269Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6863373Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6863463Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6863562Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6863761Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6863824Z res = mod(**inputs) 2025-12-04T09:48:44.6864082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6864177Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6864421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6864497Z outputs = layer_module( 2025-12-04T09:48:44.6864742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6864831Z outputs = self.rel_attn( 2025-12-04T09:48:44.6865076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6865171Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6865174Z 2025-12-04T09:48:44.6865278Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6865467Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6865531Z res = mod(**inputs) 2025-12-04T09:48:44.6865782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6865862Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6866113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6866179Z outputs = layer_module( 2025-12-04T09:48:44.6866424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6866500Z outputs = self.rel_attn( 2025-12-04T09:48:44.6866743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6866839Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6866850Z 2025-12-04T09:48:44.6866948Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6867134Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6867202Z res = mod(**inputs) 2025-12-04T09:48:44.6867446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6867525Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6867776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6867843Z outputs = layer_module( 2025-12-04T09:48:44.6868090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6868157Z outputs = self.rel_attn( 2025-12-04T09:48:44.6868396Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6868475Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6868731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6868856Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6868867Z 2025-12-04T09:48:44.6868967Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6869163Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6869265Z res = mod(**inputs) 2025-12-04T09:48:44.6869512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6869592Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6869845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6869926Z outputs = layer_module( 2025-12-04T09:48:44.6870185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6870251Z outputs = self.rel_attn( 2025-12-04T09:48:44.6870491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6870638Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6870643Z 2025-12-04T09:48:44.6870744Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6870932Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6871002Z res = mod(**inputs) 2025-12-04T09:48:44.6871246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6871333Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6871575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6871649Z outputs = layer_module( 2025-12-04T09:48:44.6871889Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6871953Z outputs = self.rel_attn( 2025-12-04T09:48:44.6872197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6872267Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6872515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6872641Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6872646Z 2025-12-04T09:48:44.6872741Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6872922Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6872987Z res = mod(**inputs) 2025-12-04T09:48:44.6873223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6873306Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6873549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6873615Z outputs = layer_module( 2025-12-04T09:48:44.6873862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6873928Z outputs = self.rel_attn( 2025-12-04T09:48:44.6874183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6874277Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6874280Z 2025-12-04T09:48:44.6874375Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6874566Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6874626Z res = mod(**inputs) 2025-12-04T09:48:44.6874866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6874982Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6875222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6875296Z outputs = layer_module( 2025-12-04T09:48:44.6875537Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6875622Z outputs = self.rel_attn( 2025-12-04T09:48:44.6875875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6875945Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6876207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6876334Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6876362Z 2025-12-04T09:48:44.6876473Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6876663Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6876724Z res = mod(**inputs) 2025-12-04T09:48:44.6876965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6877052Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6877290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6877361Z outputs = layer_module( 2025-12-04T09:48:44.6877601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6877668Z outputs = self.rel_attn( 2025-12-04T09:48:44.6877920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6878008Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6878271Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6878388Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6878391Z 2025-12-04T09:48:44.6878491Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6878686Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6878750Z res = mod(**inputs) 2025-12-04T09:48:44.6878994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6879081Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6879325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6879402Z outputs = layer_module( 2025-12-04T09:48:44.6879647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6879714Z outputs = self.rel_attn( 2025-12-04T09:48:44.6879963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6880048Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6880311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6880425Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6880428Z 2025-12-04T09:48:44.6880505Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6880592Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6880687Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6880778Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6880886Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6881078Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6881143Z res = mod(**inputs) 2025-12-04T09:48:44.6881395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6881492Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6881840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6881911Z outputs = layer_module( 2025-12-04T09:48:44.6882175Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6882259Z outputs = self.rel_attn( 2025-12-04T09:48:44.6882556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6882674Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6882678Z 2025-12-04T09:48:44.6882791Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6883015Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6883096Z res = mod(**inputs) 2025-12-04T09:48:44.6883376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6883465Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6883733Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6883803Z outputs = layer_module( 2025-12-04T09:48:44.6884104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6884173Z outputs = self.rel_attn( 2025-12-04T09:48:44.6884415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6884522Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6884526Z 2025-12-04T09:48:44.6884624Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6884822Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6884884Z res = mod(**inputs) 2025-12-04T09:48:44.6885127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6885216Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6885461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6885528Z outputs = layer_module( 2025-12-04T09:48:44.6885780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6885846Z outputs = self.rel_attn( 2025-12-04T09:48:44.6886094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6886167Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6886428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6886560Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6886563Z 2025-12-04T09:48:44.6886663Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6886856Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6886957Z res = mod(**inputs) 2025-12-04T09:48:44.6887203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6887292Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6887535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6887619Z outputs = layer_module( 2025-12-04T09:48:44.6887871Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6887938Z outputs = self.rel_attn( 2025-12-04T09:48:44.6888192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6888320Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6888339Z 2025-12-04T09:48:44.6888440Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6888635Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6888697Z res = mod(**inputs) 2025-12-04T09:48:44.6888939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6889027Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6889269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6889341Z outputs = layer_module( 2025-12-04T09:48:44.6889578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6889644Z outputs = self.rel_attn( 2025-12-04T09:48:44.6889893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6889964Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6890234Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6890359Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6890363Z 2025-12-04T09:48:44.6890461Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6890657Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6890718Z res = mod(**inputs) 2025-12-04T09:48:44.6890959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6891050Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6891297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6891371Z outputs = layer_module( 2025-12-04T09:48:44.6891610Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6891677Z outputs = self.rel_attn( 2025-12-04T09:48:44.6891923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6892018Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6892022Z 2025-12-04T09:48:44.6892127Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6892311Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6892373Z res = mod(**inputs) 2025-12-04T09:48:44.6892625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6892737Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6892981Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6893056Z outputs = layer_module( 2025-12-04T09:48:44.6893300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6893396Z outputs = self.rel_attn( 2025-12-04T09:48:44.6893640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6893709Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6893976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6894095Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6894099Z 2025-12-04T09:48:44.6894226Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6894424Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6894487Z res = mod(**inputs) 2025-12-04T09:48:44.6894739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6894820Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6895063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6895134Z outputs = layer_module( 2025-12-04T09:48:44.6895385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6895458Z outputs = self.rel_attn( 2025-12-04T09:48:44.6895699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6895788Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6896055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6896165Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6896168Z 2025-12-04T09:48:44.6896272Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6896496Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6896560Z res = mod(**inputs) 2025-12-04T09:48:44.6896826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6896905Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6897150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6897226Z outputs = layer_module( 2025-12-04T09:48:44.6897471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6897544Z outputs = self.rel_attn( 2025-12-04T09:48:44.6897784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6897870Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6898141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6898250Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6898254Z 2025-12-04T09:48:44.6898331Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6898415Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6898491Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6898604Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6898705Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6898893Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6898961Z res = mod(**inputs) 2025-12-04T09:48:44.6899202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6899306Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6899556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6899620Z outputs = layer_module( 2025-12-04T09:48:44.6899871Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6899938Z outputs = self.rel_attn( 2025-12-04T09:48:44.6900217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6900320Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6900324Z 2025-12-04T09:48:44.6900422Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6900616Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6900681Z res = mod(**inputs) 2025-12-04T09:48:44.6900923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6901009Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6901253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6901318Z outputs = layer_module( 2025-12-04T09:48:44.6901569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6901636Z outputs = self.rel_attn( 2025-12-04T09:48:44.6901883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6901978Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6901982Z 2025-12-04T09:48:44.6902081Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6902278Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6902340Z res = mod(**inputs) 2025-12-04T09:48:44.6902583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6902671Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6902916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6902993Z outputs = layer_module( 2025-12-04T09:48:44.6903235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6903301Z outputs = self.rel_attn( 2025-12-04T09:48:44.6903549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6903621Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6903884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6904011Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6904014Z 2025-12-04T09:48:44.6904116Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6904314Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6904436Z res = mod(**inputs) 2025-12-04T09:48:44.6904684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6904773Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6905044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6905131Z outputs = layer_module( 2025-12-04T09:48:44.6905376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6905442Z outputs = self.rel_attn( 2025-12-04T09:48:44.6905695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6905821Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6905841Z 2025-12-04T09:48:44.6905950Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6906136Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6906199Z res = mod(**inputs) 2025-12-04T09:48:44.6906451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6906532Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6906776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6906851Z outputs = layer_module( 2025-12-04T09:48:44.6907097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6907173Z outputs = self.rel_attn( 2025-12-04T09:48:44.6907420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6907490Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6907755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6907878Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6907882Z 2025-12-04T09:48:44.6907989Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6908178Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6908240Z res = mod(**inputs) 2025-12-04T09:48:44.6908495Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6908575Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6908820Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6908896Z outputs = layer_module( 2025-12-04T09:48:44.6909136Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6909209Z outputs = self.rel_attn( 2025-12-04T09:48:44.6909463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6909558Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6909561Z 2025-12-04T09:48:44.6909664Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6909850Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6909921Z res = mod(**inputs) 2025-12-04T09:48:44.6910164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6910279Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6910534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6910598Z outputs = layer_module( 2025-12-04T09:48:44.6910841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6910935Z outputs = self.rel_attn( 2025-12-04T09:48:44.6911179Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6911261Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6911522Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6911641Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6911646Z 2025-12-04T09:48:44.6911780Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6911971Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6912035Z res = mod(**inputs) 2025-12-04T09:48:44.6912287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6912369Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6912617Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6912683Z outputs = layer_module( 2025-12-04T09:48:44.6912925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6913000Z outputs = self.rel_attn( 2025-12-04T09:48:44.6913245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6913341Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6913603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6913711Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6913715Z 2025-12-04T09:48:44.6913823Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6914010Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6914072Z res = mod(**inputs) 2025-12-04T09:48:44.6914322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6914399Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6934040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6934165Z outputs = layer_module( 2025-12-04T09:48:44.6934459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6934542Z outputs = self.rel_attn( 2025-12-04T09:48:44.6934814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6934926Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6935199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6935319Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6935333Z 2025-12-04T09:48:44.6935424Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6935506Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6935593Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6935833Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6935945Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6936164Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6936234Z res = mod(**inputs) 2025-12-04T09:48:44.6936488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6936619Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6936872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6936955Z outputs = layer_module( 2025-12-04T09:48:44.6937205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6937280Z outputs = self.rel_attn( 2025-12-04T09:48:44.6937574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6937676Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6937680Z 2025-12-04T09:48:44.6937795Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6937995Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6938063Z res = mod(**inputs) 2025-12-04T09:48:44.6938318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6938406Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6938655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6938734Z outputs = layer_module( 2025-12-04T09:48:44.6938984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6939063Z outputs = self.rel_attn( 2025-12-04T09:48:44.6939309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6939411Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6939415Z 2025-12-04T09:48:44.6939532Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6939729Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6939795Z res = mod(**inputs) 2025-12-04T09:48:44.6940056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6940140Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6940399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6940471Z outputs = layer_module( 2025-12-04T09:48:44.6940715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6940790Z outputs = self.rel_attn( 2025-12-04T09:48:44.6941035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6941117Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6941381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6941517Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6941522Z 2025-12-04T09:48:44.6941633Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6941828Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6941928Z res = mod(**inputs) 2025-12-04T09:48:44.6942181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6942263Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6942513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6942598Z outputs = layer_module( 2025-12-04T09:48:44.6942841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6942918Z outputs = self.rel_attn( 2025-12-04T09:48:44.6943161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6943302Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6943307Z 2025-12-04T09:48:44.6943424Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6943616Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6943687Z res = mod(**inputs) 2025-12-04T09:48:44.6943933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6944014Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6944265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6944331Z outputs = layer_module( 2025-12-04T09:48:44.6944582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6944651Z outputs = self.rel_attn( 2025-12-04T09:48:44.6944894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6944979Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6945238Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6945373Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6945377Z 2025-12-04T09:48:44.6945480Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6945671Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6945743Z res = mod(**inputs) 2025-12-04T09:48:44.6945990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6946072Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6946327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6946397Z outputs = layer_module( 2025-12-04T09:48:44.6946645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6946711Z outputs = self.rel_attn( 2025-12-04T09:48:44.6946954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6947065Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6947069Z 2025-12-04T09:48:44.6947170Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6947368Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6947432Z res = mod(**inputs) 2025-12-04T09:48:44.6947680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6947785Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6948051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6948117Z outputs = layer_module( 2025-12-04T09:48:44.6948370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6948459Z outputs = self.rel_attn( 2025-12-04T09:48:44.6948708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6948779Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6949039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6949171Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6949174Z 2025-12-04T09:48:44.6949290Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6949490Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6949553Z res = mod(**inputs) 2025-12-04T09:48:44.6949798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6949888Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6950133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6950199Z outputs = layer_module( 2025-12-04T09:48:44.6950449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6950518Z outputs = self.rel_attn( 2025-12-04T09:48:44.6950772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6950867Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6951128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6951249Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6951253Z 2025-12-04T09:48:44.6951352Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6951544Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6951618Z res = mod(**inputs) 2025-12-04T09:48:44.6951863Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6951954Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6952200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6952268Z outputs = layer_module( 2025-12-04T09:48:44.6952518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6952585Z outputs = self.rel_attn( 2025-12-04T09:48:44.6952835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6952927Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6953193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6953311Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6953316Z 2025-12-04T09:48:44.6953398Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6953475Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6953561Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6953655Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6953779Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6953973Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6954039Z res = mod(**inputs) 2025-12-04T09:48:44.6954295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6954401Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6954648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6954723Z outputs = layer_module( 2025-12-04T09:48:44.6954965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6955043Z outputs = self.rel_attn( 2025-12-04T09:48:44.6955330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6955428Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6955432Z 2025-12-04T09:48:44.6955549Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6955733Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6955804Z res = mod(**inputs) 2025-12-04T09:48:44.6956046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6956124Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6956375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6956440Z outputs = layer_module( 2025-12-04T09:48:44.6956677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6956753Z outputs = self.rel_attn( 2025-12-04T09:48:44.6956987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6957092Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6957095Z 2025-12-04T09:48:44.6957193Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6957377Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6957448Z res = mod(**inputs) 2025-12-04T09:48:44.6957685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6957764Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6958009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6958076Z outputs = layer_module( 2025-12-04T09:48:44.6958320Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6958387Z outputs = self.rel_attn( 2025-12-04T09:48:44.6958624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6958706Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6958965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6959103Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6959107Z 2025-12-04T09:48:44.6959207Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6959400Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6959486Z res = mod(**inputs) 2025-12-04T09:48:44.6959750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6959834Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6960086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6960167Z outputs = layer_module( 2025-12-04T09:48:44.6960416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6960482Z outputs = self.rel_attn( 2025-12-04T09:48:44.6960722Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6960859Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6960864Z 2025-12-04T09:48:44.6960979Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6961187Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6961252Z res = mod(**inputs) 2025-12-04T09:48:44.6961504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6961696Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6961969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6962038Z outputs = layer_module( 2025-12-04T09:48:44.6962312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6962388Z outputs = self.rel_attn( 2025-12-04T09:48:44.6962664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6962755Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6963043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6963178Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6963191Z 2025-12-04T09:48:44.6963300Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6963503Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6963578Z res = mod(**inputs) 2025-12-04T09:48:44.6963836Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6963923Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6964192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6964265Z outputs = layer_module( 2025-12-04T09:48:44.6964531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6964601Z outputs = self.rel_attn( 2025-12-04T09:48:44.6964865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6964972Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6964976Z 2025-12-04T09:48:44.6965077Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6965284Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6965355Z res = mod(**inputs) 2025-12-04T09:48:44.6965599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6965709Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6965971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6966037Z outputs = layer_module( 2025-12-04T09:48:44.6966289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6966377Z outputs = self.rel_attn( 2025-12-04T09:48:44.6966618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6966694Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6966952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6967081Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6967084Z 2025-12-04T09:48:44.6967199Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6967388Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6967459Z res = mod(**inputs) 2025-12-04T09:48:44.6967703Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6967792Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6968041Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6968108Z outputs = layer_module( 2025-12-04T09:48:44.6968363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6968433Z outputs = self.rel_attn( 2025-12-04T09:48:44.6968682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6968779Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6969052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6969167Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6969170Z 2025-12-04T09:48:44.6969269Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6969461Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6969531Z res = mod(**inputs) 2025-12-04T09:48:44.6969776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6969865Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6970108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6970176Z outputs = layer_module( 2025-12-04T09:48:44.6970420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6970488Z outputs = self.rel_attn( 2025-12-04T09:48:44.6970726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6970821Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6971082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6971195Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6971199Z 2025-12-04T09:48:44.6971279Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6971358Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6971439Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6971534Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6971650Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6971851Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6971914Z res = mod(**inputs) 2025-12-04T09:48:44.6972165Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6972263Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6972506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6972580Z outputs = layer_module( 2025-12-04T09:48:44.6972822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6972896Z outputs = self.rel_attn( 2025-12-04T09:48:44.6973157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6973256Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6973259Z 2025-12-04T09:48:44.6973366Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6973557Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6973621Z res = mod(**inputs) 2025-12-04T09:48:44.6973875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6973956Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6974206Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6974272Z outputs = layer_module( 2025-12-04T09:48:44.6974514Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6974595Z outputs = self.rel_attn( 2025-12-04T09:48:44.6974836Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6974931Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6974941Z 2025-12-04T09:48:44.6975039Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6975230Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6975298Z res = mod(**inputs) 2025-12-04T09:48:44.6975542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6975623Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6975875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6975945Z outputs = layer_module( 2025-12-04T09:48:44.6976198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6976266Z outputs = self.rel_attn( 2025-12-04T09:48:44.6976512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6976592Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6976856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6976986Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6976997Z 2025-12-04T09:48:44.6977097Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6977301Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6977387Z res = mod(**inputs) 2025-12-04T09:48:44.6977658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6977738Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6977992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6978076Z outputs = layer_module( 2025-12-04T09:48:44.6978328Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6978396Z outputs = self.rel_attn( 2025-12-04T09:48:44.6978642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6978777Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6978781Z 2025-12-04T09:48:44.6978897Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6979089Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6979160Z res = mod(**inputs) 2025-12-04T09:48:44.6979403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6979490Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6979734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6979799Z outputs = layer_module( 2025-12-04T09:48:44.6980048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6980114Z outputs = self.rel_attn( 2025-12-04T09:48:44.6980359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6980434Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6980691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6980822Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6980826Z 2025-12-04T09:48:44.6980925Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6981122Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6981193Z res = mod(**inputs) 2025-12-04T09:48:44.6981436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6981524Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6981769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6981838Z outputs = layer_module( 2025-12-04T09:48:44.6982085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6982152Z outputs = self.rel_attn( 2025-12-04T09:48:44.6982402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6982496Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6982500Z 2025-12-04T09:48:44.6982599Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6982794Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6982856Z res = mod(**inputs) 2025-12-04T09:48:44.6983107Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6983211Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6983472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6983545Z outputs = layer_module( 2025-12-04T09:48:44.6983786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6983871Z outputs = self.rel_attn( 2025-12-04T09:48:44.6984119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6984187Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6984444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.6984570Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.6984573Z 2025-12-04T09:48:44.6984688Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6984888Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6984951Z res = mod(**inputs) 2025-12-04T09:48:44.6985193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6985281Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6985523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6985595Z outputs = layer_module( 2025-12-04T09:48:44.6985834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6985900Z outputs = self.rel_attn( 2025-12-04T09:48:44.6986150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6986238Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6986499Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6986613Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6986616Z 2025-12-04T09:48:44.6986715Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6986913Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6986975Z res = mod(**inputs) 2025-12-04T09:48:44.6987216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6987304Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6987546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6987621Z outputs = layer_module( 2025-12-04T09:48:44.6987862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6987929Z outputs = self.rel_attn( 2025-12-04T09:48:44.6988181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.6988267Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.6988531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.6988646Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.6988649Z 2025-12-04T09:48:44.6988727Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6988813Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6988886Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6988977Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.6989098Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6989289Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6989351Z res = mod(**inputs) 2025-12-04T09:48:44.6989606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6989702Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6989955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6990021Z outputs = layer_module( 2025-12-04T09:48:44.6990264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6990341Z outputs = self.rel_attn( 2025-12-04T09:48:44.6990598Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.6990705Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.6990709Z 2025-12-04T09:48:44.6990806Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6990993Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6991066Z res = mod(**inputs) 2025-12-04T09:48:44.6991308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6991388Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6991639Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6991704Z outputs = layer_module( 2025-12-04T09:48:44.6991950Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6992021Z outputs = self.rel_attn( 2025-12-04T09:48:44.6992260Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.6992364Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.6992367Z 2025-12-04T09:48:44.6992465Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6992660Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6992722Z res = mod(**inputs) 2025-12-04T09:48:44.6992964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6993050Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6993296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6993364Z outputs = layer_module( 2025-12-04T09:48:44.6993615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6993682Z outputs = self.rel_attn( 2025-12-04T09:48:44.6993932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6994006Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6994268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.6994400Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.6994403Z 2025-12-04T09:48:44.6994502Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6994690Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6994786Z res = mod(**inputs) 2025-12-04T09:48:44.6995052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6995140Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6995386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6995469Z outputs = layer_module( 2025-12-04T09:48:44.6995719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6995786Z outputs = self.rel_attn( 2025-12-04T09:48:44.6996035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.6996162Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.6996165Z 2025-12-04T09:48:44.6996281Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6996482Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6996548Z res = mod(**inputs) 2025-12-04T09:48:44.6996797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6996883Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6997132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6997203Z outputs = layer_module( 2025-12-04T09:48:44.6997445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6997512Z outputs = self.rel_attn( 2025-12-04T09:48:44.6997767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.6997840Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.6998102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.6998226Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.6998230Z 2025-12-04T09:48:44.6998329Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.6998529Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.6998593Z res = mod(**inputs) 2025-12-04T09:48:44.6998840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.6998927Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.6999174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.6999250Z outputs = layer_module( 2025-12-04T09:48:44.6999493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.6999559Z outputs = self.rel_attn( 2025-12-04T09:48:44.6999808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.6999904Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.6999908Z 2025-12-04T09:48:44.7000016Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.7000215Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.7000280Z res = mod(**inputs) 2025-12-04T09:48:44.7000542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.7000653Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.7000921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.7000999Z outputs = layer_module( 2025-12-04T09:48:44.7001247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.7001323Z outputs = self.rel_attn( 2025-12-04T09:48:44.7001589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.7001768Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.7002057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.7002182Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.7002187Z 2025-12-04T09:48:44.7002329Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.7002561Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.7002633Z res = mod(**inputs) 2025-12-04T09:48:44.7002925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.7003015Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.7003305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.7003388Z outputs = layer_module( 2025-12-04T09:48:44.7003668Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.7003744Z outputs = self.rel_attn( 2025-12-04T09:48:44.7003994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.7004087Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.7004362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.7004475Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.7004478Z 2025-12-04T09:48:44.7004589Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.7004784Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.7004849Z res = mod(**inputs) 2025-12-04T09:48:44.7005109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.7005190Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.7005439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.7005515Z outputs = layer_module( 2025-12-04T09:48:44.7005764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.7005839Z outputs = self.rel_attn( 2025-12-04T09:48:44.7006087Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.7006175Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.7006450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.7006559Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.7006563Z 2025-12-04T09:48:44.7006645Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.7006734Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.7006813Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.7006939Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.7007063Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.7007263Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.7007335Z res = mod(**inputs) 2025-12-04T09:48:44.7007592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.7007692Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.7007965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.7008032Z outputs = layer_module( 2025-12-04T09:48:44.7008287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.7008354Z outputs = self.rel_attn( 2025-12-04T09:48:44.7008618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 416, in forward 2025-12-04T09:48:44.7008727Z q_head_h = torch.einsum("ibh,hnd->ibnd", h, self.q) 2025-12-04T09:48:44.7008731Z 2025-12-04T09:48:44.7008832Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.7009033Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.7009099Z res = mod(**inputs) 2025-12-04T09:48:44.7009350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.7009437Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.7009686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.7009754Z outputs = layer_module( 2025-12-04T09:48:44.7010009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.7010080Z outputs = self.rel_attn( 2025-12-04T09:48:44.7010334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 417, in forward 2025-12-04T09:48:44.7010430Z k_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.k) 2025-12-04T09:48:44.7010433Z 2025-12-04T09:48:44.7010533Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.7010738Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.7010801Z res = mod(**inputs) 2025-12-04T09:48:44.7011054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.7011143Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.7011392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.7011469Z outputs = layer_module( 2025-12-04T09:48:44.7011718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.7011787Z outputs = self.rel_attn( 2025-12-04T09:48:44.7012045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.7012118Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.7012391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 263, in rel_attn_core 2025-12-04T09:48:44.7012520Z ac = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_w_bias, k_head_h) 2025-12-04T09:48:44.7012524Z 2025-12-04T09:48:44.7012625Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.7012824Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.7012909Z res = mod(**inputs) 2025-12-04T09:48:44.7013173Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.7013265Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.7013513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.7013604Z outputs = layer_module( 2025-12-04T09:48:44.7013853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.7013922Z outputs = self.rel_attn( 2025-12-04T09:48:44.7014179Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 422, in forward 2025-12-04T09:48:44.7014310Z k_head_r = torch.einsum("ibh,hnd->ibnd", r.type(self.r.dtype), self.r) 2025-12-04T09:48:44.7014313Z 2025-12-04T09:48:44.7014440Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.7014638Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.7014703Z res = mod(**inputs) 2025-12-04T09:48:44.7014963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.7015044Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.7015296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.7015370Z outputs = layer_module( 2025-12-04T09:48:44.7015619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.7015694Z outputs = self.rel_attn( 2025-12-04T09:48:44.7015945Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.7016019Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.7016296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 266, in rel_attn_core 2025-12-04T09:48:44.7016425Z bd = torch.einsum("ibnd,jbnd->bnij", q_head + self.r_r_bias, k_head_r) 2025-12-04T09:48:44.7016428Z 2025-12-04T09:48:44.7016539Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.7016735Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.7016800Z res = mod(**inputs) 2025-12-04T09:48:44.7017074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.7017159Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.7017428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.7017505Z outputs = layer_module( 2025-12-04T09:48:44.7017755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.7017832Z outputs = self.rel_attn( 2025-12-04T09:48:44.7018082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 418, in forward 2025-12-04T09:48:44.7018180Z v_head_h = torch.einsum("ibh,hnd->ibnd", cat, self.v) 2025-12-04T09:48:44.7018184Z 2025-12-04T09:48:44.7018296Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.7018493Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.7018563Z res = mod(**inputs) 2025-12-04T09:48:44.7018814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.7018897Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.7019181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.7019250Z outputs = layer_module( 2025-12-04T09:48:44.7019498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.7019573Z outputs = self.rel_attn( 2025-12-04T09:48:44.7019838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 425, in forward 2025-12-04T09:48:44.7019915Z attn_vec = self.rel_attn_core( 2025-12-04T09:48:44.7020179Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 294, in rel_attn_core 2025-12-04T09:48:44.7020300Z attn_vec = torch.einsum("bnij,jbnd->ibnd", attn_prob, v_head_h) 2025-12-04T09:48:44.7020303Z 2025-12-04T09:48:44.7020436Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.7020636Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.7020700Z res = mod(**inputs) 2025-12-04T09:48:44.7020957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.7021038Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.7021299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.7021366Z outputs = layer_module( 2025-12-04T09:48:44.7021613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.7021690Z outputs = self.rel_attn( 2025-12-04T09:48:44.7021943Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.7022038Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.7022312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.7022423Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.7022426Z 2025-12-04T09:48:44.7022535Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.7022736Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.7022800Z res = mod(**inputs) 2025-12-04T09:48:44.7023056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1616, in forward 2025-12-04T09:48:44.7023136Z transformer_outputs = self.transformer( 2025-12-04T09:48:44.7023393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1368, in forward 2025-12-04T09:48:44.7023462Z outputs = layer_module( 2025-12-04T09:48:44.7023716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 494, in forward 2025-12-04T09:48:44.7023793Z outputs = self.rel_attn( 2025-12-04T09:48:44.7024050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 440, in forward 2025-12-04T09:48:44.7024146Z output_h = self.post_attention(h, attn_vec) 2025-12-04T09:48:44.7024408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 304, in post_attention 2025-12-04T09:48:44.7024515Z attn_out = torch.einsum("ibnd,hnd->ibh", attn_vec, self.o) 2025-12-04T09:48:44.7024518Z 2025-12-04T09:48:44.7024604Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.7024682Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.7024758Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.7024841Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.7024949Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.7025031Z cudagraph partition due to non gpu ops 2025-12-04T09:48:44.7025130Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:48:44.7025319Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:48:44.7025391Z res = mod(**inputs) 2025-12-04T09:48:44.7025656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1639, in forward 2025-12-04T09:48:44.7025783Z loss = loss_fct(logits.view(-1, logits.size(-1)), labels.view(-1)) 2025-12-04T09:48:44.7025787Z 2025-12-04T09:48:57.6149590Z Compilation time (from dynamo_timed): 37.358181365 2025-12-04T09:48:57.6184557Z pass 2025-12-04T09:48:57.6185064Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:48:57.6186260Z TIMING: _recursive_pre_grad_passes:0.05896 _recursive_joint_graph_passes:1.32816 _recursive_post_grad_passes:0.67583 async_compile.wait:0.76423 code_gen:12.49239 inductor_compile:17.34088 backend_compile:31.30249 gc:0.00035 entire_frame_compile:37.35818 total_wall_time:37.35818 2025-12-04T09:48:57.6187235Z STATS: call_* op count: 820 | FakeTensorMode.__torch_dispatch__:73617 | FakeTensor.__torch_dispatch__:13337 | ProxyTorchDispatchMode.__torch_dispatch__:8773 2025-12-04T09:48:57.6187758Z Dynamo produced 1 graphs covering 820 ops with 0 graph breaks (0 unique) 2025-12-04T09:49:00.9101844Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-12-04T09:49:00.9104254Z import pynvml # type: ignore[import] 2025-12-04T09:49:04.0758688Z 2025-12-04T09:49:05.0702571Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:49:05.0705878Z loading model: 0it [00:00, ?it/s] 2025-12-04T09:49:05.0706992Z cpu eval YituTechConvBert 2025-12-04T09:49:05.9327692Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:49:06.1294750Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:49:06.3302911Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:49:23.0440517Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0440851Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0441075Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0441291Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0441500Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0441814Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0442047Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0442310Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0442541Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0442774Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0443049Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0443470Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0443843Z res = mod(**inputs) 2025-12-04T09:49:23.0444261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0444697Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0445136Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0445590Z hidden_states = self.encoder( 2025-12-04T09:49:23.0446010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0446779Z layer_outputs = layer_module( 2025-12-04T09:49:23.0447166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0447636Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0448097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0448599Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0449007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0449402Z self_outputs = self.self( 2025-12-04T09:49:23.0449800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0450336Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0450839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0451243Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0451388Z 2025-12-04T09:49:23.0451497Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0451864Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0452198Z res = mod(**inputs) 2025-12-04T09:49:23.0452573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0453063Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0453473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0453876Z hidden_states = self.encoder( 2025-12-04T09:49:23.0454270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0454679Z layer_outputs = layer_module( 2025-12-04T09:49:23.0455040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0455409Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0455828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0456253Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0456668Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0457101Z self_outputs = self.self( 2025-12-04T09:49:23.0457548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0458050Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0458547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0458999Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0459141Z 2025-12-04T09:49:23.0459250Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0459621Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0459948Z res = mod(**inputs) 2025-12-04T09:49:23.0460331Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0460806Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0461219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0461703Z hidden_states = self.encoder( 2025-12-04T09:49:23.0462110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0462516Z layer_outputs = layer_module( 2025-12-04T09:49:23.0462865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0463254Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0463673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0464089Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0464498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0464967Z self_outputs = self.self( 2025-12-04T09:49:23.0465383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0465889Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0466388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 282, in forward 2025-12-04T09:49:23.0466820Z x = self.pointwise(x) 2025-12-04T09:49:23.0466944Z 2025-12-04T09:49:23.0467069Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0467468Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0467799Z res = mod(**inputs) 2025-12-04T09:49:23.0468192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0468617Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0469038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0469478Z hidden_states = self.encoder( 2025-12-04T09:49:23.0469892Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0470304Z layer_outputs = layer_module( 2025-12-04T09:49:23.0470663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0471048Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0471496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0471940Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0472399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0472816Z self_outputs = self.self( 2025-12-04T09:49:23.0473218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0473682Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0473878Z 2025-12-04T09:49:23.0473993Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0474375Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0474709Z res = mod(**inputs) 2025-12-04T09:49:23.0475094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0475521Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0475944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0476380Z hidden_states = self.encoder( 2025-12-04T09:49:23.0476800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0477229Z layer_outputs = layer_module( 2025-12-04T09:49:23.0477605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0478010Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0478460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0478900Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0479350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0479783Z self_outputs = self.self( 2025-12-04T09:49:23.0480258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0480746Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0480931Z 2025-12-04T09:49:23.0481045Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0481490Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0481973Z res = mod(**inputs) 2025-12-04T09:49:23.0482390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0482850Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0483294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0483733Z hidden_states = self.encoder( 2025-12-04T09:49:23.0484160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0484591Z layer_outputs = layer_module( 2025-12-04T09:49:23.0484973Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0485369Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0485799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0486246Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0486682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0487108Z self_outputs = self.self( 2025-12-04T09:49:23.0487518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 379, in forward 2025-12-04T09:49:23.0488017Z conv_out_layer = torch.matmul(conv_out_layer, conv_kernel_layer) 2025-12-04T09:49:23.0488213Z 2025-12-04T09:49:23.0488310Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0488537Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0488797Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0489188Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0489542Z res = mod(**inputs) 2025-12-04T09:49:23.0489940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0490396Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0490838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0491280Z hidden_states = self.encoder( 2025-12-04T09:49:23.0491720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0492169Z layer_outputs = layer_module( 2025-12-04T09:49:23.0492547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0492939Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0493387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0493856Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0494306Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0494741Z self_outputs = self.self( 2025-12-04T09:49:23.0495174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 404, in forward 2025-12-04T09:49:23.0495669Z context_layer = torch.cat([context_layer, conv_out], 2) 2025-12-04T09:49:23.0495852Z 2025-12-04T09:49:23.0495970Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0496195Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0496422Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0496652Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0496863Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0497070Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0497282Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0497501Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0497716Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0497937Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0498157Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0498373Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0498593Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0498831Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0499047Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0499267Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0499519Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0499906Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0500247Z res = mod(**inputs) 2025-12-04T09:49:23.0500652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0501094Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0501526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0501958Z hidden_states = self.encoder( 2025-12-04T09:49:23.0502382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0502815Z layer_outputs = layer_module( 2025-12-04T09:49:23.0503180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0503553Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0503962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0504389Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0504819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0505248Z self_outputs = self.self( 2025-12-04T09:49:23.0505666Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0506185Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0506757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0507182Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0507315Z 2025-12-04T09:49:23.0507441Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0507826Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0508194Z res = mod(**inputs) 2025-12-04T09:49:23.0508602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0509041Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0509447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0509854Z hidden_states = self.encoder( 2025-12-04T09:49:23.0510271Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0510683Z layer_outputs = layer_module( 2025-12-04T09:49:23.0511056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0511447Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0511883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0512324Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0512752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0513159Z self_outputs = self.self( 2025-12-04T09:49:23.0513549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0514070Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0514606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0515052Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0515190Z 2025-12-04T09:49:23.0515305Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0515695Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0516041Z res = mod(**inputs) 2025-12-04T09:49:23.0516449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0516860Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0517278Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0517700Z hidden_states = self.encoder( 2025-12-04T09:49:23.0518137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0518558Z layer_outputs = layer_module( 2025-12-04T09:49:23.0518936Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0519325Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0519765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0520218Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0520671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0521110Z self_outputs = self.self( 2025-12-04T09:49:23.0521657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0522224Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0522774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 282, in forward 2025-12-04T09:49:23.0523240Z x = self.pointwise(x) 2025-12-04T09:49:23.0523361Z 2025-12-04T09:49:23.0523474Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0523865Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0524213Z res = mod(**inputs) 2025-12-04T09:49:23.0524610Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0525068Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0525513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0525942Z hidden_states = self.encoder( 2025-12-04T09:49:23.0526360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0526790Z layer_outputs = layer_module( 2025-12-04T09:49:23.0527164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0527546Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0527984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0528427Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0528867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0529278Z self_outputs = self.self( 2025-12-04T09:49:23.0529673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0530455Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0530652Z 2025-12-04T09:49:23.0530775Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0531147Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0531483Z res = mod(**inputs) 2025-12-04T09:49:23.0531871Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0532287Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0532716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0533128Z hidden_states = self.encoder( 2025-12-04T09:49:23.0533532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0533937Z layer_outputs = layer_module( 2025-12-04T09:49:23.0534295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0534672Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0535089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0535507Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0535928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0536338Z self_outputs = self.self( 2025-12-04T09:49:23.0536818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0537278Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0537459Z 2025-12-04T09:49:23.0537566Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0537932Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0538304Z res = mod(**inputs) 2025-12-04T09:49:23.0538693Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0539115Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0539528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0539937Z hidden_states = self.encoder( 2025-12-04T09:49:23.0540404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0540813Z layer_outputs = layer_module( 2025-12-04T09:49:23.0541172Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0541556Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0541962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0542371Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0542769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0543161Z self_outputs = self.self( 2025-12-04T09:49:23.0543547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 379, in forward 2025-12-04T09:49:23.0544013Z conv_out_layer = torch.matmul(conv_out_layer, conv_kernel_layer) 2025-12-04T09:49:23.0544204Z 2025-12-04T09:49:23.0544288Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0544512Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0544758Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0545121Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0545456Z res = mod(**inputs) 2025-12-04T09:49:23.0545846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0546259Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0546677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0547086Z hidden_states = self.encoder( 2025-12-04T09:49:23.0547492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0547895Z layer_outputs = layer_module( 2025-12-04T09:49:23.0548250Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0548620Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0549037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0549456Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0549877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0550284Z self_outputs = self.self( 2025-12-04T09:49:23.0550676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 404, in forward 2025-12-04T09:49:23.0551168Z context_layer = torch.cat([context_layer, conv_out], 2) 2025-12-04T09:49:23.0551349Z 2025-12-04T09:49:23.0551430Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0551650Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0551869Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0552097Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0552348Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0552551Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0552763Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0552974Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0553176Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0553385Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0553594Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0553800Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0554004Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0554230Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0554444Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0554646Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0554890Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0555261Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0555585Z res = mod(**inputs) 2025-12-04T09:49:23.0555974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0556395Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0556810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0557212Z hidden_states = self.encoder( 2025-12-04T09:49:23.0557617Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0558027Z layer_outputs = layer_module( 2025-12-04T09:49:23.0558374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0558773Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0559222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0559667Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0560118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0560559Z self_outputs = self.self( 2025-12-04T09:49:23.0560992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0561607Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0562157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0562613Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0562764Z 2025-12-04T09:49:23.0562881Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0563254Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0563585Z res = mod(**inputs) 2025-12-04T09:49:23.0563972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0564404Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0564815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0565930Z hidden_states = self.encoder( 2025-12-04T09:49:23.0566353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0566762Z layer_outputs = layer_module( 2025-12-04T09:49:23.0567109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0567481Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0567922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0568336Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0568757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0569167Z self_outputs = self.self( 2025-12-04T09:49:23.0569577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0570069Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0570570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0570980Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0571114Z 2025-12-04T09:49:23.0571228Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0571590Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0571935Z res = mod(**inputs) 2025-12-04T09:49:23.0572339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0572773Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0573209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0573620Z hidden_states = self.encoder( 2025-12-04T09:49:23.0574020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0574450Z layer_outputs = layer_module( 2025-12-04T09:49:23.0574821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0575463Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0575909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0576355Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0576807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0577239Z self_outputs = self.self( 2025-12-04T09:49:23.0577650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0578171Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0578708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 282, in forward 2025-12-04T09:49:23.0579140Z x = self.pointwise(x) 2025-12-04T09:49:23.0579259Z 2025-12-04T09:49:23.0579372Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0579762Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0580127Z res = mod(**inputs) 2025-12-04T09:49:23.0580547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0580998Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0581475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0581914Z hidden_states = self.encoder( 2025-12-04T09:49:23.0582345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0582797Z layer_outputs = layer_module( 2025-12-04T09:49:23.0583174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0583565Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0584007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0584447Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0584903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0585331Z self_outputs = self.self( 2025-12-04T09:49:23.0585741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0586219Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0586405Z 2025-12-04T09:49:23.0586527Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0586908Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0587258Z res = mod(**inputs) 2025-12-04T09:49:23.0587663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0588107Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0588537Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0588971Z hidden_states = self.encoder( 2025-12-04T09:49:23.0589393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0589820Z layer_outputs = layer_module( 2025-12-04T09:49:23.0590184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0590574Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0591013Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0591441Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0591880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0592304Z self_outputs = self.self( 2025-12-04T09:49:23.0592720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0593187Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0593378Z 2025-12-04T09:49:23.0593490Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0593874Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0594217Z res = mod(**inputs) 2025-12-04T09:49:23.0594595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0595015Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0595425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0595828Z hidden_states = self.encoder( 2025-12-04T09:49:23.0596279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0596687Z layer_outputs = layer_module( 2025-12-04T09:49:23.0597046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0597407Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0597853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0598303Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0598742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0599178Z self_outputs = self.self( 2025-12-04T09:49:23.0599619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 379, in forward 2025-12-04T09:49:23.0600130Z conv_out_layer = torch.matmul(conv_out_layer, conv_kernel_layer) 2025-12-04T09:49:23.0600327Z 2025-12-04T09:49:23.0600415Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0600650Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0600909Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0601293Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0601724Z res = mod(**inputs) 2025-12-04T09:49:23.0602150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0602620Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0603072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0603489Z hidden_states = self.encoder( 2025-12-04T09:49:23.0603901Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0604313Z layer_outputs = layer_module( 2025-12-04T09:49:23.0604662Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0605036Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0605450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0605861Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0606277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0606683Z self_outputs = self.self( 2025-12-04T09:49:23.0607077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 404, in forward 2025-12-04T09:49:23.0607524Z context_layer = torch.cat([context_layer, conv_out], 2) 2025-12-04T09:49:23.0607699Z 2025-12-04T09:49:23.0607782Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0608002Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0608211Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0608426Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0608639Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0608850Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0609054Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0609262Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0609476Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0609673Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0609876Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0610081Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0610278Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0610523Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0610730Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0610929Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0611168Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0611528Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0611872Z res = mod(**inputs) 2025-12-04T09:49:23.0612241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0612655Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0613067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0613461Z hidden_states = self.encoder( 2025-12-04T09:49:23.0613875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0614271Z layer_outputs = layer_module( 2025-12-04T09:49:23.0614621Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0614978Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0615381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0615790Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0616196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0616581Z self_outputs = self.self( 2025-12-04T09:49:23.0616966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0617452Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0617928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0618331Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0618466Z 2025-12-04T09:49:23.0618569Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0618934Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0619257Z res = mod(**inputs) 2025-12-04T09:49:23.0619641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0620070Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0620477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0620879Z hidden_states = self.encoder( 2025-12-04T09:49:23.0621279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0621686Z layer_outputs = layer_module( 2025-12-04T09:49:23.0622041Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0622403Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0622802Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0623206Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0623602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0623999Z self_outputs = self.self( 2025-12-04T09:49:23.0624398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0624899Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0625375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0625776Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0625924Z 2025-12-04T09:49:23.0626035Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0626388Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0626715Z res = mod(**inputs) 2025-12-04T09:49:23.0627077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0627469Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0627872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0628268Z hidden_states = self.encoder( 2025-12-04T09:49:23.0628658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0629046Z layer_outputs = layer_module( 2025-12-04T09:49:23.0629379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0629734Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0630239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0630642Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0631044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0631442Z self_outputs = self.self( 2025-12-04T09:49:23.0631830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0632308Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0632783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 282, in forward 2025-12-04T09:49:23.0633173Z x = self.pointwise(x) 2025-12-04T09:49:23.0633284Z 2025-12-04T09:49:23.0633394Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0633738Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0634057Z res = mod(**inputs) 2025-12-04T09:49:23.0634425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0634822Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0635219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0635608Z hidden_states = self.encoder( 2025-12-04T09:49:23.0635989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0636369Z layer_outputs = layer_module( 2025-12-04T09:49:23.0636711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0637060Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0637445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0637841Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0638282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0638688Z self_outputs = self.self( 2025-12-04T09:49:23.0639058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0639528Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0639723Z 2025-12-04T09:49:23.0639836Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0640198Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0640517Z res = mod(**inputs) 2025-12-04T09:49:23.0640901Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0641319Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0641809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0642232Z hidden_states = self.encoder( 2025-12-04T09:49:23.0642655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0643085Z layer_outputs = layer_module( 2025-12-04T09:49:23.0643452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0643843Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0644261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0644678Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0645088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0645497Z self_outputs = self.self( 2025-12-04T09:49:23.0645887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0646335Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0646519Z 2025-12-04T09:49:23.0646625Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0646987Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0647316Z res = mod(**inputs) 2025-12-04T09:49:23.0647691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0648113Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0648527Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0648941Z hidden_states = self.encoder( 2025-12-04T09:49:23.0649333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0649743Z layer_outputs = layer_module( 2025-12-04T09:49:23.0650097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0650461Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0650876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0651293Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0651706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0652112Z self_outputs = self.self( 2025-12-04T09:49:23.0652509Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 379, in forward 2025-12-04T09:49:23.0653028Z conv_out_layer = torch.matmul(conv_out_layer, conv_kernel_layer) 2025-12-04T09:49:23.0653211Z 2025-12-04T09:49:23.0653302Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0653520Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0653765Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0654131Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0654480Z res = mod(**inputs) 2025-12-04T09:49:23.0654856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0655265Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0655663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0656072Z hidden_states = self.encoder( 2025-12-04T09:49:23.0656469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0656863Z layer_outputs = layer_module( 2025-12-04T09:49:23.0657200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0657564Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0657967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0658372Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0658772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0659173Z self_outputs = self.self( 2025-12-04T09:49:23.0659566Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 404, in forward 2025-12-04T09:49:23.0660008Z context_layer = torch.cat([context_layer, conv_out], 2) 2025-12-04T09:49:23.0660180Z 2025-12-04T09:49:23.0660260Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0660472Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0660681Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0660879Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0661083Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0661292Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0661487Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0661691Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0661899Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0662099Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0662308Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0662516Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0662724Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0662926Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0663136Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0663345Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0663576Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0663944Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0664274Z res = mod(**inputs) 2025-12-04T09:49:23.0664651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0665070Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0665485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0665894Z hidden_states = self.encoder( 2025-12-04T09:49:23.0666348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0666781Z layer_outputs = layer_module( 2025-12-04T09:49:23.0667138Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0667501Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0667913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0668360Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0668780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0669185Z self_outputs = self.self( 2025-12-04T09:49:23.0669588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0670094Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0670585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0670994Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0671133Z 2025-12-04T09:49:23.0671241Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0671612Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0671935Z res = mod(**inputs) 2025-12-04T09:49:23.0672317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0672735Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0673150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0673560Z hidden_states = self.encoder( 2025-12-04T09:49:23.0673972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0674391Z layer_outputs = layer_module( 2025-12-04T09:49:23.0674788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0675151Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0675565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0675988Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0676394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0676805Z self_outputs = self.self( 2025-12-04T09:49:23.0677214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0677717Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0678214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0678636Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0678780Z 2025-12-04T09:49:23.0678889Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0679281Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0679610Z res = mod(**inputs) 2025-12-04T09:49:23.0680000Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0680438Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0680917Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0681351Z hidden_states = self.encoder( 2025-12-04T09:49:23.0681855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0682310Z layer_outputs = layer_module( 2025-12-04T09:49:23.0682722Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0683121Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0683573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0683991Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0684417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0684827Z self_outputs = self.self( 2025-12-04T09:49:23.0685225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0685740Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0686268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 282, in forward 2025-12-04T09:49:23.0686697Z x = self.pointwise(x) 2025-12-04T09:49:23.0686818Z 2025-12-04T09:49:23.0686938Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0687319Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0687668Z res = mod(**inputs) 2025-12-04T09:49:23.0688054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0688501Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0688933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0689366Z hidden_states = self.encoder( 2025-12-04T09:49:23.0689789Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0690212Z layer_outputs = layer_module( 2025-12-04T09:49:23.0690590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0690982Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0691417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0691852Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0692285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0692691Z self_outputs = self.self( 2025-12-04T09:49:23.0693077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0693533Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0693723Z 2025-12-04T09:49:23.0693839Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0694226Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0694567Z res = mod(**inputs) 2025-12-04T09:49:23.0694973Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0695414Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0695872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0696313Z hidden_states = self.encoder( 2025-12-04T09:49:23.0696737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0697164Z layer_outputs = layer_module( 2025-12-04T09:49:23.0697530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0697942Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0698381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0698817Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0699248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0699706Z self_outputs = self.self( 2025-12-04T09:49:23.0700126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0700607Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0700792Z 2025-12-04T09:49:23.0700904Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0701296Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0701645Z res = mod(**inputs) 2025-12-04T09:49:23.0702045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0702488Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0702939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0703382Z hidden_states = self.encoder( 2025-12-04T09:49:23.0703807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0704245Z layer_outputs = layer_module( 2025-12-04T09:49:23.0704623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0705015Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0705450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0705897Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0706339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0706767Z self_outputs = self.self( 2025-12-04T09:49:23.0707192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 379, in forward 2025-12-04T09:49:23.0707704Z conv_out_layer = torch.matmul(conv_out_layer, conv_kernel_layer) 2025-12-04T09:49:23.0707898Z 2025-12-04T09:49:23.0707997Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0708225Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0708493Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0708893Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0709245Z res = mod(**inputs) 2025-12-04T09:49:23.0709650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0710075Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0710495Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0710923Z hidden_states = self.encoder( 2025-12-04T09:49:23.0711340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0711750Z layer_outputs = layer_module( 2025-12-04T09:49:23.0712099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0712519Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0712948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0713369Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0713778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0714189Z self_outputs = self.self( 2025-12-04T09:49:23.0714601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 404, in forward 2025-12-04T09:49:23.0715059Z context_layer = torch.cat([context_layer, conv_out], 2) 2025-12-04T09:49:23.0715227Z 2025-12-04T09:49:23.0715312Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0715536Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0715752Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0715956Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0716169Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0716383Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0716587Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0716795Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0717004Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0717211Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0717414Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0717621Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0717832Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0718034Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0718244Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0718452Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0718685Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0719054Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0719386Z res = mod(**inputs) 2025-12-04T09:49:23.0719774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0720189Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0720609Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0721022Z hidden_states = self.encoder( 2025-12-04T09:49:23.0721422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0721911Z layer_outputs = layer_module( 2025-12-04T09:49:23.0722289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0722681Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0723113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0723561Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0723976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0724387Z self_outputs = self.self( 2025-12-04T09:49:23.0724784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0725323Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0725820Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0726224Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0726366Z 2025-12-04T09:49:23.0726474Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0726858Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0727186Z res = mod(**inputs) 2025-12-04T09:49:23.0727562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0727983Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0728412Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0728834Z hidden_states = self.encoder( 2025-12-04T09:49:23.0729233Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0729645Z layer_outputs = layer_module( 2025-12-04T09:49:23.0730001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0730520Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0730943Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0731369Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0731790Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0732196Z self_outputs = self.self( 2025-12-04T09:49:23.0732609Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0733107Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0733605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0734018Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0734160Z 2025-12-04T09:49:23.0734267Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0734643Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0734967Z res = mod(**inputs) 2025-12-04T09:49:23.0735362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0735793Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0736212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0736613Z hidden_states = self.encoder( 2025-12-04T09:49:23.0737011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0737418Z layer_outputs = layer_module( 2025-12-04T09:49:23.0737762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0738127Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0738543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0738962Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0739378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0739858Z self_outputs = self.self( 2025-12-04T09:49:23.0740261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0740757Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0741248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 282, in forward 2025-12-04T09:49:23.0741686Z x = self.pointwise(x) 2025-12-04T09:49:23.0741798Z 2025-12-04T09:49:23.0741913Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0742270Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0742598Z res = mod(**inputs) 2025-12-04T09:49:23.0743005Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0743430Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0743849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0744252Z hidden_states = self.encoder( 2025-12-04T09:49:23.0744653Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0745072Z layer_outputs = layer_module( 2025-12-04T09:49:23.0745409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0745765Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0746165Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0746563Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0746977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0747370Z self_outputs = self.self( 2025-12-04T09:49:23.0747755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0748190Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0748369Z 2025-12-04T09:49:23.0748474Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0748832Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0749153Z res = mod(**inputs) 2025-12-04T09:49:23.0749519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0749926Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0750345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0750745Z hidden_states = self.encoder( 2025-12-04T09:49:23.0751151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0751556Z layer_outputs = layer_module( 2025-12-04T09:49:23.0751912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0752277Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0752690Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0753102Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0753506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0753935Z self_outputs = self.self( 2025-12-04T09:49:23.0754357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0754811Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0754986Z 2025-12-04T09:49:23.0755094Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0755492Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0755819Z res = mod(**inputs) 2025-12-04T09:49:23.0756200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0756609Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0757024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0757449Z hidden_states = self.encoder( 2025-12-04T09:49:23.0757847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0758257Z layer_outputs = layer_module( 2025-12-04T09:49:23.0758614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0758990Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0759399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0759816Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0760231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0760641Z self_outputs = self.self( 2025-12-04T09:49:23.0761034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 379, in forward 2025-12-04T09:49:23.0761525Z conv_out_layer = torch.matmul(conv_out_layer, conv_kernel_layer) 2025-12-04T09:49:23.0761807Z 2025-12-04T09:49:23.0761912Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0762149Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0762426Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0762835Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0763190Z res = mod(**inputs) 2025-12-04T09:49:23.0763596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0764049Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0764473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0764879Z hidden_states = self.encoder( 2025-12-04T09:49:23.0765290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0765701Z layer_outputs = layer_module( 2025-12-04T09:49:23.0766058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0766426Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0766839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0767262Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0767676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0768078Z self_outputs = self.self( 2025-12-04T09:49:23.0768490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 404, in forward 2025-12-04T09:49:23.0768957Z context_layer = torch.cat([context_layer, conv_out], 2) 2025-12-04T09:49:23.0769126Z 2025-12-04T09:49:23.0769207Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0769428Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0769642Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0769876Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0770079Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0770289Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0770502Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0770711Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0770930Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0771150Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0771368Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0771598Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0771813Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0772016Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0772225Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0772435Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0772674Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0773038Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0773373Z res = mod(**inputs) 2025-12-04T09:49:23.0773759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0774175Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0774594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0775005Z hidden_states = self.encoder( 2025-12-04T09:49:23.0775416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0775819Z layer_outputs = layer_module( 2025-12-04T09:49:23.0776176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0776552Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0776965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0777394Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0777817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0778226Z self_outputs = self.self( 2025-12-04T09:49:23.0778621Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0779128Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0779633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0780046Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0780181Z 2025-12-04T09:49:23.0780290Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0780663Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0781035Z res = mod(**inputs) 2025-12-04T09:49:23.0781449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0781908Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0782355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0782803Z hidden_states = self.encoder( 2025-12-04T09:49:23.0783201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0783609Z layer_outputs = layer_module( 2025-12-04T09:49:23.0783972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0784391Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0784831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0785278Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0785722Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0786155Z self_outputs = self.self( 2025-12-04T09:49:23.0786603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0787141Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0787675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0788119Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0788267Z 2025-12-04T09:49:23.0788380Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0788781Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0789140Z res = mod(**inputs) 2025-12-04T09:49:23.0789552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0789997Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0790448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0790889Z hidden_states = self.encoder( 2025-12-04T09:49:23.0791324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0791803Z layer_outputs = layer_module( 2025-12-04T09:49:23.0792186Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0792574Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0793029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0793567Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0794039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0794458Z self_outputs = self.self( 2025-12-04T09:49:23.0794855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0795347Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0795840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 282, in forward 2025-12-04T09:49:23.0796251Z x = self.pointwise(x) 2025-12-04T09:49:23.0796371Z 2025-12-04T09:49:23.0796477Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0796850Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0797173Z res = mod(**inputs) 2025-12-04T09:49:23.0797558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0798012Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0798425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0798899Z hidden_states = self.encoder( 2025-12-04T09:49:23.0799330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0799792Z layer_outputs = layer_module( 2025-12-04T09:49:23.0800164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0800538Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0800980Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0801431Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0801981Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0802445Z self_outputs = self.self( 2025-12-04T09:49:23.0802878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0803366Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0803564Z 2025-12-04T09:49:23.0803679Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0804066Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0804395Z res = mod(**inputs) 2025-12-04T09:49:23.0804769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0805185Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0805601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0806007Z hidden_states = self.encoder( 2025-12-04T09:49:23.0806397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0806799Z layer_outputs = layer_module( 2025-12-04T09:49:23.0807150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0807508Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0807925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0808339Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0808750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0809149Z self_outputs = self.self( 2025-12-04T09:49:23.0809543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0809997Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0810172Z 2025-12-04T09:49:23.0810283Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0810642Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0810973Z res = mod(**inputs) 2025-12-04T09:49:23.0811357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0811775Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0812180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0812595Z hidden_states = self.encoder( 2025-12-04T09:49:23.0813011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0813401Z layer_outputs = layer_module( 2025-12-04T09:49:23.0813749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0814124Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0814518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0814933Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0815338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0815733Z self_outputs = self.self( 2025-12-04T09:49:23.0816129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 379, in forward 2025-12-04T09:49:23.0816587Z conv_out_layer = torch.matmul(conv_out_layer, conv_kernel_layer) 2025-12-04T09:49:23.0816772Z 2025-12-04T09:49:23.0816855Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0817071Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0817306Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0817669Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0817992Z res = mod(**inputs) 2025-12-04T09:49:23.0818361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0818772Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0819181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0819583Z hidden_states = self.encoder( 2025-12-04T09:49:23.0819971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0820370Z layer_outputs = layer_module( 2025-12-04T09:49:23.0820719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0821076Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0821479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0821886Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0822289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0822678Z self_outputs = self.self( 2025-12-04T09:49:23.0823066Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 404, in forward 2025-12-04T09:49:23.0823511Z context_layer = torch.cat([context_layer, conv_out], 2) 2025-12-04T09:49:23.0823676Z 2025-12-04T09:49:23.0823764Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0823973Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0824185Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0824394Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0824593Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0824801Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0825004Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0825203Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0825411Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0825617Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0825816Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0826044Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0826262Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0826467Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0826663Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0826867Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0827101Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0827457Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0827808Z res = mod(**inputs) 2025-12-04T09:49:23.0828194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0828621Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0829037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0829449Z hidden_states = self.encoder( 2025-12-04T09:49:23.0829871Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0830388Z layer_outputs = layer_module( 2025-12-04T09:49:23.0830752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0831128Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0831550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0831960Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0832380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0832455Z self_outputs = self.self( 2025-12-04T09:49:23.0832737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0832906Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0833186Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0833265Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0833269Z 2025-12-04T09:49:23.0833378Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0833591Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0833660Z res = mod(**inputs) 2025-12-04T09:49:23.0833932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0834023Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0834296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0834381Z hidden_states = self.encoder( 2025-12-04T09:49:23.0834651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0834722Z layer_outputs = layer_module( 2025-12-04T09:49:23.0834955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0835038Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0835316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0835398Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0835670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0835748Z self_outputs = self.self( 2025-12-04T09:49:23.0836089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0836252Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0836530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0836632Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0836635Z 2025-12-04T09:49:23.0836749Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0836955Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0837022Z res = mod(**inputs) 2025-12-04T09:49:23.0837314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0837420Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0837707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0837781Z hidden_states = self.encoder( 2025-12-04T09:49:23.0838051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0838131Z layer_outputs = layer_module( 2025-12-04T09:49:23.0838358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0838434Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0838712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0838793Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0839072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0839146Z self_outputs = self.self( 2025-12-04T09:49:23.0839414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0839583Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0839854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 282, in forward 2025-12-04T09:49:23.0839932Z x = self.pointwise(x) 2025-12-04T09:49:23.0839936Z 2025-12-04T09:49:23.0840042Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0840244Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0840317Z res = mod(**inputs) 2025-12-04T09:49:23.0840588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0840681Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0840953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0841025Z hidden_states = self.encoder( 2025-12-04T09:49:23.0841324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0841402Z layer_outputs = layer_module( 2025-12-04T09:49:23.0841711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0841816Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0842123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0842224Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0842570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0842648Z self_outputs = self.self( 2025-12-04T09:49:23.0842942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0843071Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0843094Z 2025-12-04T09:49:23.0843218Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0843432Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0843499Z res = mod(**inputs) 2025-12-04T09:49:23.0843776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0843860Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0844157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0844244Z hidden_states = self.encoder( 2025-12-04T09:49:23.0844512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0844591Z layer_outputs = layer_module( 2025-12-04T09:49:23.0844817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0844897Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0845172Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0845254Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0845522Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0845604Z self_outputs = self.self( 2025-12-04T09:49:23.0845868Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0845993Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0845996Z 2025-12-04T09:49:23.0846101Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0846306Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0846378Z res = mod(**inputs) 2025-12-04T09:49:23.0846640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0846727Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0846990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0847064Z hidden_states = self.encoder( 2025-12-04T09:49:23.0847349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0847419Z layer_outputs = layer_module( 2025-12-04T09:49:23.0847635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0847720Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0847980Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0848067Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0848343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0848413Z self_outputs = self.self( 2025-12-04T09:49:23.0848730Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 379, in forward 2025-12-04T09:49:23.0848859Z conv_out_layer = torch.matmul(conv_out_layer, conv_kernel_layer) 2025-12-04T09:49:23.0848863Z 2025-12-04T09:49:23.0848952Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0849030Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0849132Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0849352Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0849420Z res = mod(**inputs) 2025-12-04T09:49:23.0849680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0849768Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0850046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0850129Z hidden_states = self.encoder( 2025-12-04T09:49:23.0850392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0850462Z layer_outputs = layer_module( 2025-12-04T09:49:23.0850688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0850769Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0851038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0851119Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0851381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0851457Z self_outputs = self.self( 2025-12-04T09:49:23.0851723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 404, in forward 2025-12-04T09:49:23.0851834Z context_layer = torch.cat([context_layer, conv_out], 2) 2025-12-04T09:49:23.0851838Z 2025-12-04T09:49:23.0851924Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0852001Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0852083Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0852160Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0852234Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0852318Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0852391Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0852464Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0852545Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0852620Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0852694Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0852779Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0852856Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0852936Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0853010Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0853084Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0853193Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0853389Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0853454Z res = mod(**inputs) 2025-12-04T09:49:23.0853722Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0853803Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0854079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0854170Z hidden_states = self.encoder( 2025-12-04T09:49:23.0854451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0854530Z layer_outputs = layer_module( 2025-12-04T09:49:23.0854747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0854826Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0855111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0855193Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0855465Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0855536Z self_outputs = self.self( 2025-12-04T09:49:23.0855848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0856016Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0856279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0856361Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0856365Z 2025-12-04T09:49:23.0856468Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0856661Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0856733Z res = mod(**inputs) 2025-12-04T09:49:23.0856997Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0857080Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0857354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0857427Z hidden_states = self.encoder( 2025-12-04T09:49:23.0857702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0857774Z layer_outputs = layer_module( 2025-12-04T09:49:23.0858000Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0858100Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0858362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0858452Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0858712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0858785Z self_outputs = self.self( 2025-12-04T09:49:23.0859052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0859209Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0859474Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0859559Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0859562Z 2025-12-04T09:49:23.0859663Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0859872Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0859935Z res = mod(**inputs) 2025-12-04T09:49:23.0860189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0860294Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0860564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0860643Z hidden_states = self.encoder( 2025-12-04T09:49:23.0860897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0860982Z layer_outputs = layer_module( 2025-12-04T09:49:23.0861205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0861280Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0861534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0861621Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0861899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0861977Z self_outputs = self.self( 2025-12-04T09:49:23.0862241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0862394Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0862670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 282, in forward 2025-12-04T09:49:23.0862739Z x = self.pointwise(x) 2025-12-04T09:49:23.0862743Z 2025-12-04T09:49:23.0862852Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0863048Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0863113Z res = mod(**inputs) 2025-12-04T09:49:23.0863385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0863466Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0863731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0863811Z hidden_states = self.encoder( 2025-12-04T09:49:23.0864076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0864162Z layer_outputs = layer_module( 2025-12-04T09:49:23.0864379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0864456Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0864718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0864801Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0865068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0865137Z self_outputs = self.self( 2025-12-04T09:49:23.0865393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0865517Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0865520Z 2025-12-04T09:49:23.0865622Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0865809Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0865879Z res = mod(**inputs) 2025-12-04T09:49:23.0866141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0866230Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0866529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0866604Z hidden_states = self.encoder( 2025-12-04T09:49:23.0866878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0866949Z layer_outputs = layer_module( 2025-12-04T09:49:23.0867207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0867287Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0867555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0867648Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0867944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0868019Z self_outputs = self.self( 2025-12-04T09:49:23.0868299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0868416Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0868420Z 2025-12-04T09:49:23.0868531Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0868731Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0868798Z res = mod(**inputs) 2025-12-04T09:49:23.0869074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0869156Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0869432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0869506Z hidden_states = self.encoder( 2025-12-04T09:49:23.0869775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0869854Z layer_outputs = layer_module( 2025-12-04T09:49:23.0870080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0870160Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0870435Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0870517Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0870806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0870875Z self_outputs = self.self( 2025-12-04T09:49:23.0871142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 379, in forward 2025-12-04T09:49:23.0871276Z conv_out_layer = torch.matmul(conv_out_layer, conv_kernel_layer) 2025-12-04T09:49:23.0871280Z 2025-12-04T09:49:23.0871358Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0871441Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0871545Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0871739Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0871810Z res = mod(**inputs) 2025-12-04T09:49:23.0872070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0872150Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0872420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0872524Z hidden_states = self.encoder( 2025-12-04T09:49:23.0872794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0872867Z layer_outputs = layer_module( 2025-12-04T09:49:23.0873084Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0873188Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0873452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0873534Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0873820Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0873892Z self_outputs = self.self( 2025-12-04T09:49:23.0874192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 404, in forward 2025-12-04T09:49:23.0874307Z context_layer = torch.cat([context_layer, conv_out], 2) 2025-12-04T09:49:23.0874311Z 2025-12-04T09:49:23.0874391Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0874478Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0874558Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0874635Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0874720Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0874796Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0874879Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0874955Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0875031Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0875116Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0875193Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0875272Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0875358Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0875434Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0875509Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0875591Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0875696Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0875906Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0875972Z res = mod(**inputs) 2025-12-04T09:49:23.0876244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0876334Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0876608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0876683Z hidden_states = self.encoder( 2025-12-04T09:49:23.0876970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0877042Z layer_outputs = layer_module( 2025-12-04T09:49:23.0877276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0877357Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0877629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0877721Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0877996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0878074Z self_outputs = self.self( 2025-12-04T09:49:23.0878369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0878546Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0878827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0878905Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0878925Z 2025-12-04T09:49:23.0879039Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0879241Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0879306Z res = mod(**inputs) 2025-12-04T09:49:23.0879587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0879673Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0879962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0880048Z hidden_states = self.encoder( 2025-12-04T09:49:23.0880321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0880400Z layer_outputs = layer_module( 2025-12-04T09:49:23.0880625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0880705Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0880983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0881064Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0881335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0881416Z self_outputs = self.self( 2025-12-04T09:49:23.0881756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0881941Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0882257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0882344Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0882348Z 2025-12-04T09:49:23.0882483Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0882706Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0882786Z res = mod(**inputs) 2025-12-04T09:49:23.0883076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0883167Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0883460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0883538Z hidden_states = self.encoder( 2025-12-04T09:49:23.0883823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0883911Z layer_outputs = layer_module( 2025-12-04T09:49:23.0884149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0884241Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0884544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0884632Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0884949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0885046Z self_outputs = self.self( 2025-12-04T09:49:23.0885343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0885513Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0885817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 282, in forward 2025-12-04T09:49:23.0885900Z x = self.pointwise(x) 2025-12-04T09:49:23.0885904Z 2025-12-04T09:49:23.0886014Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0886231Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0886301Z res = mod(**inputs) 2025-12-04T09:49:23.0886620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0886718Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0887009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0887086Z hidden_states = self.encoder( 2025-12-04T09:49:23.0887398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0887475Z layer_outputs = layer_module( 2025-12-04T09:49:23.0887725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0887809Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0888122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0888220Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0888513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0888588Z self_outputs = self.self( 2025-12-04T09:49:23.0888887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0889016Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0889020Z 2025-12-04T09:49:23.0889139Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0889350Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0889420Z res = mod(**inputs) 2025-12-04T09:49:23.0889736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0889826Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0890126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0890203Z hidden_states = self.encoder( 2025-12-04T09:49:23.0890492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0890580Z layer_outputs = layer_module( 2025-12-04T09:49:23.0890820Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0890903Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0891212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0891299Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0891626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0891730Z self_outputs = self.self( 2025-12-04T09:49:23.0892003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0892130Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0892134Z 2025-12-04T09:49:23.0892267Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0892470Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0892537Z res = mod(**inputs) 2025-12-04T09:49:23.0892805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0892899Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0893184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0893261Z hidden_states = self.encoder( 2025-12-04T09:49:23.0893538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0893611Z layer_outputs = layer_module( 2025-12-04T09:49:23.0893840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0893920Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0894207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0894305Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0894600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0894685Z self_outputs = self.self( 2025-12-04T09:49:23.0894982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 379, in forward 2025-12-04T09:49:23.0895121Z conv_out_layer = torch.matmul(conv_out_layer, conv_kernel_layer) 2025-12-04T09:49:23.0895125Z 2025-12-04T09:49:23.0895219Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0895304Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0895419Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0895644Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0895715Z res = mod(**inputs) 2025-12-04T09:49:23.0896017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0896106Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0896402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0896490Z hidden_states = self.encoder( 2025-12-04T09:49:23.0896789Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0896873Z layer_outputs = layer_module( 2025-12-04T09:49:23.0897118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0897203Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0897506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0897595Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0897890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0897996Z self_outputs = self.self( 2025-12-04T09:49:23.0898302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 404, in forward 2025-12-04T09:49:23.0898431Z context_layer = torch.cat([context_layer, conv_out], 2) 2025-12-04T09:49:23.0898435Z 2025-12-04T09:49:23.0898519Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0898602Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0898715Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0898798Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0898879Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0898967Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0899047Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0899134Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0899215Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0899297Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0899401Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0899489Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0899571Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0899659Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0899738Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0899818Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0899934Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0900145Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0900222Z res = mod(**inputs) 2025-12-04T09:49:23.0900510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0900598Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0900891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0900972Z hidden_states = self.encoder( 2025-12-04T09:49:23.0901262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0901339Z layer_outputs = layer_module( 2025-12-04T09:49:23.0901548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0901632Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0901884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0901963Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0902226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0902295Z self_outputs = self.self( 2025-12-04T09:49:23.0902556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0902719Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0902981Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0903064Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0903069Z 2025-12-04T09:49:23.0903170Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0903363Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0903437Z res = mod(**inputs) 2025-12-04T09:49:23.0903697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0903795Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0904087Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0904160Z hidden_states = self.encoder( 2025-12-04T09:49:23.0904422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0904490Z layer_outputs = layer_module( 2025-12-04T09:49:23.0904726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0904803Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0905061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0905150Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0905425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0905499Z self_outputs = self.self( 2025-12-04T09:49:23.0905770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0905925Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0906196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0906273Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0906277Z 2025-12-04T09:49:23.0906384Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0906589Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0906653Z res = mod(**inputs) 2025-12-04T09:49:23.0906925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0907008Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0907281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0907359Z hidden_states = self.encoder( 2025-12-04T09:49:23.0907612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0907681Z layer_outputs = layer_module( 2025-12-04T09:49:23.0907901Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0907976Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0908237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0908316Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0908581Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0908658Z self_outputs = self.self( 2025-12-04T09:49:23.0908919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0909087Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0909344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 282, in forward 2025-12-04T09:49:23.0909412Z x = self.pointwise(x) 2025-12-04T09:49:23.0909415Z 2025-12-04T09:49:23.0909523Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0909712Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0909776Z res = mod(**inputs) 2025-12-04T09:49:23.0910063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0910161Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0910424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0910493Z hidden_states = self.encoder( 2025-12-04T09:49:23.0910764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0910840Z layer_outputs = layer_module( 2025-12-04T09:49:23.0911056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0911139Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0911404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0911501Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0911766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0911833Z self_outputs = self.self( 2025-12-04T09:49:23.0912087Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0912208Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0912212Z 2025-12-04T09:49:23.0912311Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0912505Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0912567Z res = mod(**inputs) 2025-12-04T09:49:23.0912827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0912915Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0913171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0913247Z hidden_states = self.encoder( 2025-12-04T09:49:23.0913503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0913572Z layer_outputs = layer_module( 2025-12-04T09:49:23.0913797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0913873Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0914126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0914212Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0914469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0914547Z self_outputs = self.self( 2025-12-04T09:49:23.0914800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0914912Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0914916Z 2025-12-04T09:49:23.0915023Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0915210Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0915280Z res = mod(**inputs) 2025-12-04T09:49:23.0915533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0915614Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0915890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0915984Z hidden_states = self.encoder( 2025-12-04T09:49:23.0916241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0916316Z layer_outputs = layer_module( 2025-12-04T09:49:23.0916527Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0916626Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0916882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0916961Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0917228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0917313Z self_outputs = self.self( 2025-12-04T09:49:23.0917575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 379, in forward 2025-12-04T09:49:23.0917706Z conv_out_layer = torch.matmul(conv_out_layer, conv_kernel_layer) 2025-12-04T09:49:23.0917709Z 2025-12-04T09:49:23.0917785Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0917870Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0917969Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0918157Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0918227Z res = mod(**inputs) 2025-12-04T09:49:23.0918482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0918568Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0918825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0918897Z hidden_states = self.encoder( 2025-12-04T09:49:23.0919159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0919227Z layer_outputs = layer_module( 2025-12-04T09:49:23.0919437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0919521Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0919777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0919864Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0920124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0920196Z self_outputs = self.self( 2025-12-04T09:49:23.0920466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 404, in forward 2025-12-04T09:49:23.0920577Z context_layer = torch.cat([context_layer, conv_out], 2) 2025-12-04T09:49:23.0920580Z 2025-12-04T09:49:23.0920665Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0920742Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0920818Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0920902Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0920977Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0921050Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0921134Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0921208Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0921280Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0921362Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0921460Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0921629Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0921713Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0921790Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0921872Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0921957Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0922072Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0922323Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0922396Z res = mod(**inputs) 2025-12-04T09:49:23.0922708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0922806Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0923116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0923209Z hidden_states = self.encoder( 2025-12-04T09:49:23.0923502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0923580Z layer_outputs = layer_module( 2025-12-04T09:49:23.0923829Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0923915Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0924204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0924286Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0924549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0924627Z self_outputs = self.self( 2025-12-04T09:49:23.0924896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0925054Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0925326Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0925401Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0925406Z 2025-12-04T09:49:23.0925517Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0925713Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0925777Z res = mod(**inputs) 2025-12-04T09:49:23.0926050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0926131Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0926406Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0926477Z hidden_states = self.encoder( 2025-12-04T09:49:23.0926741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0926820Z layer_outputs = layer_module( 2025-12-04T09:49:23.0927041Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0927120Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0927393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0927473Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0927745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0927851Z self_outputs = self.self( 2025-12-04T09:49:23.0928127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0928289Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0928553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 281, in forward 2025-12-04T09:49:23.0928656Z x = self.depthwise(hidden_states) 2025-12-04T09:49:23.0928659Z 2025-12-04T09:49:23.0928761Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0928956Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0929027Z res = mod(**inputs) 2025-12-04T09:49:23.0929307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0929391Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0929659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0929730Z hidden_states = self.encoder( 2025-12-04T09:49:23.0930003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0930073Z layer_outputs = layer_module( 2025-12-04T09:49:23.0930437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0930528Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0930792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0930882Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0931157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0931229Z self_outputs = self.self( 2025-12-04T09:49:23.0931503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 346, in forward 2025-12-04T09:49:23.0931661Z mixed_key_conv_attn_layer = self.key_conv_attn_layer(hidden_states.transpose(1, 2)) 2025-12-04T09:49:23.0931927Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 282, in forward 2025-12-04T09:49:23.0932005Z x = self.pointwise(x) 2025-12-04T09:49:23.0932009Z 2025-12-04T09:49:23.0932112Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0932315Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0932380Z res = mod(**inputs) 2025-12-04T09:49:23.0932648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0932740Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0933003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0933082Z hidden_states = self.encoder( 2025-12-04T09:49:23.0933352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0933422Z layer_outputs = layer_module( 2025-12-04T09:49:23.0933648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0933725Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0933987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0934138Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0934400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0934478Z self_outputs = self.self( 2025-12-04T09:49:23.0934739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0934877Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0934881Z 2025-12-04T09:49:23.0934991Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0935183Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0935256Z res = mod(**inputs) 2025-12-04T09:49:23.0935518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0935624Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0935899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0935971Z hidden_states = self.encoder( 2025-12-04T09:49:23.0936232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0936315Z layer_outputs = layer_module( 2025-12-04T09:49:23.0936529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0936612Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0936873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0936953Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0937226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0937295Z self_outputs = self.self( 2025-12-04T09:49:23.0937571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 361, in forward 2025-12-04T09:49:23.0937683Z conv_kernel_layer = self.conv_kernel_layer(conv_attn_layer) 2025-12-04T09:49:23.0937688Z 2025-12-04T09:49:23.0937786Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0937982Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0938044Z res = mod(**inputs) 2025-12-04T09:49:23.0938296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0938382Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0938637Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0938714Z hidden_states = self.encoder( 2025-12-04T09:49:23.0938969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0939036Z layer_outputs = layer_module( 2025-12-04T09:49:23.0939255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0939330Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0939589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0939668Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0939924Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0940030Z self_outputs = self.self( 2025-12-04T09:49:23.0940301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 379, in forward 2025-12-04T09:49:23.0940428Z conv_out_layer = torch.matmul(conv_out_layer, conv_kernel_layer) 2025-12-04T09:49:23.0940438Z 2025-12-04T09:49:23.0940515Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0940591Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0940711Z cudagraph partition due to non gpu ops. Found from : 2025-12-04T09:49:23.0940906Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/huggingface.py", line 553, in forward_pass 2025-12-04T09:49:23.0940967Z res = mod(**inputs) 2025-12-04T09:49:23.0941238Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 924, in forward 2025-12-04T09:49:23.0941317Z generator_hidden_states = self.convbert( 2025-12-04T09:49:23.0941598Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 852, in forward 2025-12-04T09:49:23.0941678Z hidden_states = self.encoder( 2025-12-04T09:49:23.0941932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 624, in forward 2025-12-04T09:49:23.0942006Z layer_outputs = layer_module( 2025-12-04T09:49:23.0942216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-12-04T09:49:23.0942292Z return super().__call__(*args, **kwargs) 2025-12-04T09:49:23.0942553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 560, in forward 2025-12-04T09:49:23.0942630Z self_attention_outputs = self.attention( 2025-12-04T09:49:23.0942891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 463, in forward 2025-12-04T09:49:23.0942960Z self_outputs = self.self( 2025-12-04T09:49:23.0943215Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/convbert/modeling_convbert.py", line 404, in forward 2025-12-04T09:49:23.0943329Z context_layer = torch.cat([context_layer, conv_out], 2) 2025-12-04T09:49:23.0943333Z 2025-12-04T09:49:23.0943408Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0943482Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0943565Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0943640Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0943719Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0943792Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0943864Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0943944Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0944018Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0944090Z cudagraph partition due to non gpu ops 2025-12-04T09:49:23.0944172Z cudagraph partition due to non gpu ops 2025-12-04T09:49:33.5218165Z Compilation time (from dynamo_timed): 26.568310292 2025-12-04T09:49:33.5247433Z pass 2025-12-04T09:49:33.5248035Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-12-04T09:49:33.5249016Z TIMING: _recursive_pre_grad_passes:0.30843 _recursive_joint_graph_passes:0.95567 _recursive_post_grad_passes:0.15167 async_compile.wait:0.60172 code_gen:10.39054 inductor_compile:13.17292 backend_compile:22.66958 gc:0.00025 entire_frame_compile:26.56831 total_wall_time:26.56831 2025-12-04T09:49:33.5249962Z STATS: call_* op count: 636 | FakeTensorMode.__torch_dispatch__:40767 | FakeTensor.__torch_dispatch__:5254 | ProxyTorchDispatchMode.__torch_dispatch__:6904 2025-12-04T09:49:33.5250516Z Dynamo produced 1 graphs covering 636 ops with 0 graph breaks (0 unique) 2025-12-04T09:49:35.4356889Z accuracy pass_rate=92.59% 2025-12-04T09:49:35.4359095Z calls_captured gmean=0.00x mean=582.852x 2025-12-04T09:49:35.4363759Z unique_graphs gmean=0.00x mean=1.185x 2025-12-04T09:49:35.4368652Z graph_breaks gmean=0.00x mean=0.222x 2025-12-04T09:49:35.4372784Z unique_graph_breaks gmean=0.00x mean=0.074x 2025-12-04T09:49:35.4378292Z autograd_captures gmean=0.00x mean=0.000x 2025-12-04T09:49:35.4382796Z autograd_compiles gmean=0.00x mean=0.000x 2025-12-04T09:49:35.4386994Z cudagraph_skips gmean=0.00x mean=1.111x 2025-12-04T09:49:35.4391006Z compilation_latency mean=21.586 seconds 2025-12-04T09:49:36.1940025Z + python benchmarks/dynamo/check_accuracy.py --actual /var/lib/jenkins/workspace/test/test-reports/inference_huggingface.csv --expected benchmarks/dynamo/ci_expected_accuracy/cpu_inductor_amp_freezing_huggingface_inference.csv 2025-12-04T09:49:36.4585183Z AlbertForMaskedLM PASS 2025-12-04T09:49:36.4586960Z AllenaiLongformerBase PASS 2025-12-04T09:49:36.4587297Z BartForCausalLM PASS 2025-12-04T09:49:36.4587959Z BertForMaskedLM PASS 2025-12-04T09:49:36.4596097Z BlenderbotForCausalLM XFAIL 2025-12-04T09:49:36.4596366Z DebertaV2ForMaskedLM XFAIL 2025-12-04T09:49:36.4596597Z DistilBertForMaskedLM PASS 2025-12-04T09:49:36.4596812Z DistillGPT2 PASS 2025-12-04T09:49:36.4601847Z ElectraForCausalLM PASS 2025-12-04T09:49:36.4602703Z GPT2ForSequenceClassification PASS 2025-12-04T09:49:36.4603138Z GoogleFnet PASS 2025-12-04T09:49:36.4603387Z LayoutLMForMaskedLM PASS 2025-12-04T09:49:36.4611466Z M2M100ForConditionalGeneration PASS 2025-12-04T09:49:36.4611881Z MBartForCausalLM PASS 2025-12-04T09:49:36.4621946Z MT5ForConditionalGeneration PASS 2025-12-04T09:49:36.4622375Z MegatronBertForCausalLM PASS 2025-12-04T09:49:36.4624394Z MobileBertForMaskedLM PASS 2025-12-04T09:49:36.4624681Z OPTForCausalLM PASS 2025-12-04T09:49:36.4636190Z PLBartForCausalLM PASS 2025-12-04T09:49:36.4636469Z PegasusForCausalLM PASS 2025-12-04T09:49:36.4636684Z RobertaForCausalLM PASS 2025-12-04T09:49:36.4636890Z T5ForConditionalGeneration PASS 2025-12-04T09:49:36.4645298Z T5Small PASS 2025-12-04T09:49:36.4645561Z TrOCRForCausalLM PASS 2025-12-04T09:49:36.4647842Z XGLMForCausalLM PASS 2025-12-04T09:49:36.4648265Z XLNetLMHeadModel PASS 2025-12-04T09:49:36.4655582Z YituTechConvBert PASS 2025-12-04T09:49:36.5109124Z + python benchmarks/dynamo/check_graph_breaks.py --actual /var/lib/jenkins/workspace/test/test-reports/inference_huggingface.csv --expected benchmarks/dynamo/ci_expected_accuracy/cpu_inductor_amp_freezing_huggingface_inference.csv 2025-12-04T09:49:36.7707640Z AlbertForMaskedLM PASS 2025-12-04T09:49:36.7709289Z AllenaiLongformerBase PASS 2025-12-04T09:49:36.7709552Z BartForCausalLM PASS 2025-12-04T09:49:36.7714027Z BertForMaskedLM PASS 2025-12-04T09:49:36.7714302Z BlenderbotForCausalLM PASS 2025-12-04T09:49:36.7717180Z DebertaV2ForMaskedLM PASS 2025-12-04T09:49:36.7717653Z DistilBertForMaskedLM PASS 2025-12-04T09:49:36.7718644Z DistillGPT2 PASS 2025-12-04T09:49:36.7719305Z ElectraForCausalLM PASS 2025-12-04T09:49:36.7728242Z GPT2ForSequenceClassification PASS 2025-12-04T09:49:36.7733132Z GoogleFnet PASS 2025-12-04T09:49:36.7737840Z LayoutLMForMaskedLM PASS 2025-12-04T09:49:36.7742656Z M2M100ForConditionalGeneration PASS 2025-12-04T09:49:36.7746962Z MBartForCausalLM PASS 2025-12-04T09:49:36.7751984Z MT5ForConditionalGeneration PASS 2025-12-04T09:49:36.7752378Z MegatronBertForCausalLM PASS 2025-12-04T09:49:36.7752711Z MobileBertForMaskedLM PASS 2025-12-04T09:49:36.7753470Z OPTForCausalLM PASS 2025-12-04T09:49:36.7759145Z PLBartForCausalLM PASS 2025-12-04T09:49:36.7759449Z PegasusForCausalLM PASS 2025-12-04T09:49:36.7761919Z RobertaForCausalLM PASS 2025-12-04T09:49:36.7762212Z T5ForConditionalGeneration PASS 2025-12-04T09:49:36.7767946Z T5Small PASS 2025-12-04T09:49:36.7773952Z TrOCRForCausalLM PASS 2025-12-04T09:49:36.7774444Z XGLMForCausalLM PASS_BUT_FLAKY 2025-12-04T09:49:36.7774701Z XLNetLMHeadModel PASS 2025-12-04T09:49:36.7784141Z YituTechConvBert PASS 2025-12-04T09:49:36.8276755Z + sccache_epilogue 2025-12-04T09:49:36.8277397Z + echo '::group::Sccache Compilation Log' 2025-12-04T09:49:36.8277914Z ##[group]Sccache Compilation Log 2025-12-04T09:49:36.8278162Z + echo '=================== sccache compilation log ===================' 2025-12-04T09:49:36.8278736Z =================== sccache compilation log =================== 2025-12-04T09:49:36.8279126Z + python /var/lib/jenkins/workspace/.ci/pytorch/print_sccache_log.py /var/lib/jenkins/sccache_error.log 2025-12-04T09:49:36.8498590Z + echo '=========== If your build fails, please take a look at the log above for possible reasons ===========' 2025-12-04T09:49:36.8503196Z =========== If your build fails, please take a look at the log above for possible reasons =========== 2025-12-04T09:49:36.8507195Z + sccache --show-stats 2025-12-04T09:49:36.8532127Z Compile requests 285 2025-12-04T09:49:36.8532429Z Compile requests executed 0 2025-12-04T09:49:36.8533278Z Cache hits 0 2025-12-04T09:49:36.8535990Z Cache misses 0 2025-12-04T09:49:36.8536207Z Cache hits rate - 2025-12-04T09:49:36.8536507Z Cache timeouts 0 2025-12-04T09:49:36.8542605Z Cache read errors 0 2025-12-04T09:49:36.8547117Z Forced recaches 0 2025-12-04T09:49:36.8551716Z Cache write errors 0 2025-12-04T09:49:36.8557612Z Cache errors 0 2025-12-04T09:49:36.8558014Z Compilations 0 2025-12-04T09:49:36.8558226Z Compilation failures 0 2025-12-04T09:49:36.8558528Z Non-cacheable compilations 0 2025-12-04T09:49:36.8558757Z Non-cacheable calls 25 2025-12-04T09:49:36.8559320Z Non-compilation calls 260 2025-12-04T09:49:36.8559606Z Unsupported compiler calls 0 2025-12-04T09:49:36.8559819Z Average cache write 0.000 s 2025-12-04T09:49:36.8560036Z Average compiler 0.000 s 2025-12-04T09:49:36.8560240Z Average cache read hit 0.000 s 2025-12-04T09:49:36.8560446Z Failed distributed compilations 0 2025-12-04T09:49:36.8560576Z 2025-12-04T09:49:36.8560647Z Non-cacheable reasons: 2025-12-04T09:49:36.8560852Z -E 25 2025-12-04T09:49:36.8560987Z 2025-12-04T09:49:36.8561152Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-12-04T09:49:36.8561467Z Version (client) 0.10.0 2025-12-04T09:49:36.8561848Z + sccache --stop-server 2025-12-04T09:49:36.8562044Z Stopping sccache server... 2025-12-04T09:49:36.8562443Z Compile requests 285 2025-12-04T09:49:36.8562797Z Compile requests executed 0 2025-12-04T09:49:36.8563097Z Cache hits 0 2025-12-04T09:49:36.8563735Z Cache misses 0 2025-12-04T09:49:36.8564038Z Cache hits rate - 2025-12-04T09:49:36.8564277Z Cache timeouts 0 2025-12-04T09:49:36.8564505Z Cache read errors 0 2025-12-04T09:49:36.8564740Z Forced recaches 0 2025-12-04T09:49:36.8564960Z Cache write errors 0 2025-12-04T09:49:36.8565202Z Cache errors 0 2025-12-04T09:49:36.8565701Z Compilations 0 2025-12-04T09:49:36.8565927Z Compilation failures 0 2025-12-04T09:49:36.8566160Z Non-cacheable compilations 0 2025-12-04T09:49:36.8566395Z Non-cacheable calls 25 2025-12-04T09:49:36.8566619Z Non-compilation calls 260 2025-12-04T09:49:36.8566855Z Unsupported compiler calls 0 2025-12-04T09:49:36.8567152Z Average cache write 0.000 s 2025-12-04T09:49:36.8567380Z Average compiler 0.000 s 2025-12-04T09:49:36.8567616Z Average cache read hit 0.000 s 2025-12-04T09:49:36.8567853Z Failed distributed compilations 0 2025-12-04T09:49:36.8568013Z 2025-12-04T09:49:36.8568100Z Non-cacheable reasons: 2025-12-04T09:49:36.8568295Z -E 25 2025-12-04T09:49:36.8568446Z 2025-12-04T09:49:36.8568623Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-12-04T09:49:36.8568987Z Version (client) 0.10.0 2025-12-04T09:49:36.8569250Z + echo ::endgroup:: 2025-12-04T09:49:36.8569678Z ##[endgroup] 2025-12-04T09:49:36.8569861Z + cleanup_workspace 2025-12-04T09:49:36.8570212Z + echo 'sudo may print the following warning message that can be ignored. The chown command will still run.' 2025-12-04T09:49:36.8570687Z sudo may print the following warning message that can be ignored. The chown command will still run. 2025-12-04T09:49:36.8571068Z + echo ' sudo: setrlimit(RLIMIT_STACK): Operation not permitted' 2025-12-04T09:49:36.8571369Z sudo: setrlimit(RLIMIT_STACK): Operation not permitted 2025-12-04T09:49:36.8571719Z + echo 'For more details refer to https://github.com/sudo-project/sudo/issues/42' 2025-12-04T09:49:36.8572062Z For more details refer to https://github.com/sudo-project/sudo/issues/42 2025-12-04T09:49:36.8572357Z + sudo chown -R 1000 /var/lib/jenkins/workspace 2025-12-04T09:49:37.2826635Z ##[group]Run pytorch/test-infra/.github/actions/upload-benchmark-results@main 2025-12-04T09:49:37.2826953Z with: 2025-12-04T09:49:37.2827133Z benchmark-results-dir: test/test-reports 2025-12-04T09:49:37.2827331Z dry-run: false 2025-12-04T09:49:37.2827497Z schema-version: v3 2025-12-04T09:49:37.2827847Z github-token: *** 2025-12-04T09:49:37.2828000Z env: 2025-12-04T09:49:37.2828141Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:37.2828320Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:37.2828611Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:37.2828899Z ##[endgroup] 2025-12-04T09:49:37.2842191Z ##[group]Run set -eux 2025-12-04T09:49:37.2842409Z set -eux 2025-12-04T09:49:37.2842584Z  2025-12-04T09:49:37.2842729Z if [[ -n "" ]]; then 2025-12-04T09:49:37.2842911Z  source "" 2025-12-04T09:49:37.2843081Z fi 2025-12-04T09:49:37.2843336Z python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-12-04T09:49:37.2843585Z  2025-12-04T09:49:37.2843733Z DEVICE_NAME="" 2025-12-04T09:49:37.2843903Z DEVICE_TYPE="" 2025-12-04T09:49:37.2844061Z  2025-12-04T09:49:37.2844211Z if command -v nvidia-smi; then 2025-12-04T09:49:37.2844482Z  # NB: I'm using PyTorch here to get the device name, however, it needs to 2025-12-04T09:49:37.2844812Z  # install the correct version of PyTorch manually for now. Any PyTorch 2025-12-04T09:49:37.2845117Z  # version is fine, I just use 2.7.1 to satify PYPIDEP linter 2025-12-04T09:49:37.2845378Z  python3 -mpip install torch==2.7.1 2025-12-04T09:49:37.2845597Z elif command -v rocminfo; then 2025-12-04T09:49:37.2845863Z  # NB: Installing torch on ROCm runner with pip here causes CI to fail 2025-12-04T09:49:37.2846187Z  # with a memoryview is too large error only on MI300 runners. Is pip 2025-12-04T09:49:37.2846508Z  # version on ROCm runner there too old? As a workaround, let's use the 2025-12-04T09:49:37.2846857Z  # GPU device name coming from rocminfo instead 2025-12-04T09:49:37.2847068Z  DEVICE_NAME=rocm 2025-12-04T09:49:37.2847358Z  DEVICE_TYPE=$(rocminfo | grep "Marketing Name" | tail -n1 | awk -F':' '{print $2}' | xargs) 2025-12-04T09:49:37.2847648Z fi 2025-12-04T09:49:37.2847796Z  2025-12-04T09:49:37.2847973Z echo "DEVICE_NAME=$DEVICE_NAME" >> $GITHUB_ENV 2025-12-04T09:49:37.2848271Z echo "DEVICE_TYPE=$DEVICE_TYPE" >> $GITHUB_ENV 2025-12-04T09:49:37.2855772Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:37.2856000Z env: 2025-12-04T09:49:37.2856159Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:37.2856340Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:37.2856612Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:37.2856913Z ##[endgroup] 2025-12-04T09:49:37.2889519Z + [[ -n '' ]] 2025-12-04T09:49:37.2889977Z + python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-12-04T09:49:37.4590441Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T09:49:38.2107001Z Collecting boto3==1.35.33 2025-12-04T09:49:38.2431846Z Downloading boto3-1.35.33-py3-none-any.whl (139 kB) 2025-12-04T09:49:38.4642325Z Collecting psutil==7.0.0 2025-12-04T09:49:38.4677039Z Downloading psutil-7.0.0-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (277 kB) 2025-12-04T09:49:38.4928914Z Collecting pynvml==12.0.0 2025-12-04T09:49:38.4959778Z Downloading pynvml-12.0.0-py3-none-any.whl (26 kB) 2025-12-04T09:49:38.5324545Z Collecting s3transfer<0.11.0,>=0.10.0 2025-12-04T09:49:38.5358600Z Downloading s3transfer-0.10.4-py3-none-any.whl (83 kB) 2025-12-04T09:49:39.3313823Z Collecting botocore<1.36.0,>=1.35.33 2025-12-04T09:49:39.3350634Z Downloading botocore-1.35.99-py3-none-any.whl (13.3 MB) 2025-12-04T09:49:39.4137254Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.33) (0.10.0) 2025-12-04T09:49:39.4450569Z Collecting nvidia-ml-py<13.0.0a0,>=12.0.0 2025-12-04T09:49:39.4482947Z Downloading nvidia_ml_py-12.575.51-py3-none-any.whl (47 kB) 2025-12-04T09:49:39.4556107Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.25.10) 2025-12-04T09:49:39.4561036Z Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (2.8.1) 2025-12-04T09:49:39.5652111Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil<3.0.0,>=2.1->botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.15.0) 2025-12-04T09:49:39.6617906Z Installing collected packages: botocore, s3transfer, nvidia-ml-py, pynvml, psutil, boto3 2025-12-04T09:49:39.9972892Z Attempting uninstall: nvidia-ml-py 2025-12-04T09:49:39.9975343Z Found existing installation: nvidia-ml-py 11.525.84 2025-12-04T09:49:39.9983689Z Uninstalling nvidia-ml-py-11.525.84: 2025-12-04T09:49:40.0110485Z Successfully uninstalled nvidia-ml-py-11.525.84 2025-12-04T09:49:40.0618632Z Attempting uninstall: psutil 2025-12-04T09:49:40.0621804Z Found existing installation: psutil 5.9.8 2025-12-04T09:49:40.0664098Z Uninstalling psutil-5.9.8: 2025-12-04T09:49:40.0670285Z Successfully uninstalled psutil-5.9.8 2025-12-04T09:49:40.1969608Z Successfully installed boto3-1.35.33 botocore-1.35.99 nvidia-ml-py-12.575.51 psutil-7.0.0 pynvml-12.0.0 s3transfer-0.10.4 2025-12-04T09:49:40.2946827Z + DEVICE_NAME= 2025-12-04T09:49:40.2948513Z + DEVICE_TYPE= 2025-12-04T09:49:40.2948835Z + command -v nvidia-smi 2025-12-04T09:49:40.2954199Z + command -v rocminfo 2025-12-04T09:49:40.2956154Z + echo DEVICE_NAME= 2025-12-04T09:49:40.2956469Z + echo DEVICE_TYPE= 2025-12-04T09:49:40.2973023Z ##[group]Run set -eux 2025-12-04T09:49:40.2973212Z set -eux 2025-12-04T09:49:40.2973431Z  2025-12-04T09:49:40.2973600Z if [[ -z "${GITHUB_TOKEN}" ]]; then 2025-12-04T09:49:40.2973824Z  echo "Missing github-token input" 2025-12-04T09:49:40.2974024Z  exit 1 2025-12-04T09:49:40.2974180Z fi 2025-12-04T09:49:40.2979883Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:40.2980115Z env: 2025-12-04T09:49:40.2980274Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:40.2980502Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:40.2980783Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:40.2981072Z DEVICE_NAME: 2025-12-04T09:49:40.2981222Z DEVICE_TYPE: 2025-12-04T09:49:40.2981565Z GITHUB_TOKEN: *** 2025-12-04T09:49:40.2981725Z ##[endgroup] 2025-12-04T09:49:40.3000249Z + [[ -z *** ]] 2025-12-04T09:49:40.3034712Z ##[group]Run pytorch/test-infra/.github/actions/get-workflow-job-id@main 2025-12-04T09:49:40.3034979Z with: 2025-12-04T09:49:40.3035242Z github-token: *** 2025-12-04T09:49:40.3035400Z env: 2025-12-04T09:49:40.3035547Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:40.3035723Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:40.3035989Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:40.3036279Z DEVICE_NAME: 2025-12-04T09:49:40.3036433Z DEVICE_TYPE: 2025-12-04T09:49:40.3036577Z ##[endgroup] 2025-12-04T09:49:40.3046872Z ##[group]Run set -eux 2025-12-04T09:49:40.3047098Z set -eux 2025-12-04T09:49:40.3047288Z  2025-12-04T09:49:40.3047656Z python3 "${GITHUB_ACTION_PATH}/../../scripts/get_workflow_job_id.py" "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-12-04T09:49:40.3051497Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:40.3051724Z env: 2025-12-04T09:49:40.3051878Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:40.3052052Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:40.3052435Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:40.3052737Z DEVICE_NAME: 2025-12-04T09:49:40.3052892Z DEVICE_TYPE: 2025-12-04T09:49:40.3053288Z GITHUB_TOKEN: *** 2025-12-04T09:49:40.3053451Z ##[endgroup] 2025-12-04T09:49:40.3076135Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/get-workflow-job-id/../../scripts/get_workflow_job_id.py 19923066595 i-0c3ca63174fa64465 2025-12-04T09:49:42.4701315Z setting job-id=57118563344 2025-12-04T09:49:42.4703732Z setting job-name=periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx) 2025-12-04T09:49:42.4816770Z ##[group]Run set -eux 2025-12-04T09:49:42.4816965Z set -eux 2025-12-04T09:49:42.4817117Z  2025-12-04T09:49:42.4817258Z if [[ -n "" ]]; then 2025-12-04T09:49:42.4817437Z  source "" 2025-12-04T09:49:42.4817594Z fi 2025-12-04T09:49:42.4817753Z  2025-12-04T09:49:42.4817987Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_metadata.py" \ 2025-12-04T09:49:42.4818298Z  --schema-version "${SCHEMA_VERSION}" \ 2025-12-04T09:49:42.4818515Z  --repo "${REPO}" \ 2025-12-04T09:49:42.4818704Z  --head-branch "${HEAD_BRANCH}" \ 2025-12-04T09:49:42.4818913Z  --head-sha "${HEAD_SHA}" \ 2025-12-04T09:49:42.4819125Z  --workflow-id "${WORKFLOW_RUN_ID}" \ 2025-12-04T09:49:42.4819344Z  --run-attempt "${RUN_ATTEMPT}" \ 2025-12-04T09:49:42.4819540Z  --job-id "${JOB_ID}" \ 2025-12-04T09:49:42.4819735Z  --job-name "${JOB_NAME}" 2025-12-04T09:49:42.4825107Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:42.4825336Z env: 2025-12-04T09:49:42.4825492Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:42.4825671Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:42.4825961Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:42.4826270Z DEVICE_NAME: 2025-12-04T09:49:42.4826520Z DEVICE_TYPE: 2025-12-04T09:49:42.4826669Z SCHEMA_VERSION: v3 2025-12-04T09:49:42.4826844Z REPO: pytorch/pytorch 2025-12-04T09:49:42.4827023Z HEAD_BRANCH: refs/heads/main 2025-12-04T09:49:42.4827243Z HEAD_SHA: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:49:42.4827465Z WORKFLOW_RUN_ID: 19923066595 2025-12-04T09:49:42.4827642Z RUN_ATTEMPT: 1 2025-12-04T09:49:42.4827802Z JOB_ID: 57118563344 2025-12-04T09:49:42.4828206Z JOB_NAME: periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx) 2025-12-04T09:49:42.4828576Z ##[endgroup] 2025-12-04T09:49:42.4853358Z + [[ -n '' ]] 2025-12-04T09:49:42.4858860Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_metadata.py --schema-version v3 --repo pytorch/pytorch --head-branch refs/heads/main --head-sha ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 --workflow-id 19923066595 --run-attempt 1 --job-id 57118563344 --job-name 'periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx)' 2025-12-04T09:49:42.5098054Z ##[group]Run set -eux 2025-12-04T09:49:42.5098248Z set -eux 2025-12-04T09:49:42.5098405Z  2025-12-04T09:49:42.5098550Z if [[ -n "" ]]; then 2025-12-04T09:49:42.5098729Z  source "" 2025-12-04T09:49:42.5098886Z fi 2025-12-04T09:49:42.5099031Z  2025-12-04T09:49:42.5099279Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_runners_info.py" 2025-12-04T09:49:42.5103308Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:42.5103537Z env: 2025-12-04T09:49:42.5103683Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:42.5103857Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:42.5104250Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:42.5104540Z DEVICE_NAME: 2025-12-04T09:49:42.5104702Z DEVICE_TYPE: 2025-12-04T09:49:42.5104854Z ##[endgroup] 2025-12-04T09:49:42.5125100Z + [[ -n '' ]] 2025-12-04T09:49:42.5125679Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_runners_info.py 2025-12-04T09:49:42.5428656Z INFO:root:Fail to import torch to get the device name 2025-12-04T09:49:42.5512858Z ##[group]Run set -eux 2025-12-04T09:49:42.5513044Z set -eux 2025-12-04T09:49:42.5513199Z  2025-12-04T09:49:42.5513368Z # TODO (huydhn): Implement this part 2025-12-04T09:49:42.5513605Z echo "dependencies={}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:49:42.5517626Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:42.5517871Z env: 2025-12-04T09:49:42.5518018Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:42.5518198Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:42.5518572Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:42.5518874Z DEVICE_NAME: 2025-12-04T09:49:42.5519023Z DEVICE_TYPE: 2025-12-04T09:49:42.5519180Z ##[endgroup] 2025-12-04T09:49:42.5541960Z + echo 'dependencies={}' 2025-12-04T09:49:42.5556804Z ##[group]Run set -eux 2025-12-04T09:49:42.5556993Z set -eux 2025-12-04T09:49:42.5557146Z  2025-12-04T09:49:42.5557299Z if [[ -n "" ]]; then 2025-12-04T09:49:42.5557465Z  source "" 2025-12-04T09:49:42.5557626Z fi 2025-12-04T09:49:42.5557769Z  2025-12-04T09:49:42.5557936Z if [[ ! -d "${BENCHMARK_RESULTS_DIR}" ]]; then 2025-12-04T09:49:42.5558202Z  echo "${BENCHMARK_RESULTS_DIR} does not exist, skipping" 2025-12-04T09:49:42.5558495Z  # We don't want the job to fail if the directory doesn't exist 2025-12-04T09:49:42.5558732Z  exit 0 2025-12-04T09:49:42.5558875Z fi 2025-12-04T09:49:42.5559025Z  2025-12-04T09:49:42.5559186Z if [[ "${DRY_RUN}" == "true" ]]; then 2025-12-04T09:49:42.5559523Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-12-04T09:49:42.5559857Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-12-04T09:49:42.5560116Z  --metadata "${BENCHMARK_METADATA}" \ 2025-12-04T09:49:42.5560336Z  --runners "${RUNNER_INFO}" \ 2025-12-04T09:49:42.5560547Z  --dependencies "${DEPENDENCIES}" \ 2025-12-04T09:49:42.5560799Z  --dry-run 2025-12-04T09:49:42.5560964Z else 2025-12-04T09:49:42.5561196Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-12-04T09:49:42.5561638Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-12-04T09:49:42.5561934Z  --metadata "${BENCHMARK_METADATA}" \ 2025-12-04T09:49:42.5562184Z  --runners "${RUNNER_INFO}" \ 2025-12-04T09:49:42.5562428Z  --dependencies "${DEPENDENCIES}" 2025-12-04T09:49:42.5562657Z fi 2025-12-04T09:49:42.5566276Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:42.5566512Z env: 2025-12-04T09:49:42.5566667Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:42.5566841Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:42.5567128Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:42.5567427Z DEVICE_NAME: 2025-12-04T09:49:42.5567581Z DEVICE_TYPE: 2025-12-04T09:49:42.5567747Z BENCHMARK_RESULTS_DIR: test/test-reports 2025-12-04T09:49:42.5567947Z DRY_RUN: false 2025-12-04T09:49:42.5568860Z BENCHMARK_METADATA: {"timestamp": 1764841782, "schema_version": "v3", "name": "periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32", "workflow_id": 19923066595, "run_attempt": 1, "job_id": 57118563344} 2025-12-04T09:49:42.5569881Z RUNNER_INFO: [{"cpu_info": "x86_64", "cpu_count": 32, "avail_mem_in_gb": 123, "extra_info": {"hostname": "ip-10-0-45-203.ec2.internal"}, "name": "", "type": ""}] 2025-12-04T09:49:42.5570240Z DEPENDENCIES: {} 2025-12-04T09:49:42.5570403Z ##[endgroup] 2025-12-04T09:49:42.5592194Z + [[ -n '' ]] 2025-12-04T09:49:42.5597040Z + [[ ! -d test/test-reports ]] 2025-12-04T09:49:42.5598726Z + [[ false == \t\r\u\e ]] 2025-12-04T09:49:42.5600460Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/upload_benchmark_results.py --benchmark-results-dir test/test-reports --metadata '{"timestamp": 1764841782, "schema_version": "v3", "name": "periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32", "workflow_id": 19923066595, "run_attempt": 1, "job_id": 57118563344}' --runners '[{"cpu_info": "x86_64", "cpu_count": 32, "avail_mem_in_gb": 123, "extra_info": {"hostname": "ip-10-0-45-203.ec2.internal"}, "name": "", "type": ""}]' --dependencies '{}' 2025-12-04T09:49:42.6669269Z INFO:root:Upload test/test-reports/inference_huggingface.json to s3://ossci-benchmarks/v3/pytorch/pytorch/19923066595/57118563344/inference_huggingface.json 2025-12-04T09:49:42.6922433Z INFO:botocore.credentials:Found credentials from IAM Role: gh-ci-github-action-runners-runner-role 2025-12-04T09:49:42.8848896Z ##[group]Run cat test/**/*_toprint.log || true 2025-12-04T09:49:42.8849161Z cat test/**/*_toprint.log || true 2025-12-04T09:49:42.8853537Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:42.8853782Z env: 2025-12-04T09:49:42.8853937Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:42.8854119Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:42.8854406Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:42.8854708Z DEVICE_NAME: 2025-12-04T09:49:42.8854935Z DEVICE_TYPE: 2025-12-04T09:49:42.8855090Z ##[endgroup] 2025-12-04T09:49:42.8925599Z cat: 'test/**/*_toprint.log': No such file or directory 2025-12-04T09:49:42.8974592Z ##[group]Run kill "$MONITOR_SCRIPT_PID" 2025-12-04T09:49:42.8974837Z kill "$MONITOR_SCRIPT_PID" 2025-12-04T09:49:42.8978888Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:42.8979123Z env: 2025-12-04T09:49:42.8979354Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:42.8979537Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:42.8979824Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:42.8980137Z DEVICE_NAME: 2025-12-04T09:49:42.8980299Z DEVICE_TYPE: 2025-12-04T09:49:42.8980466Z MONITOR_SCRIPT_PID: 52410 2025-12-04T09:49:42.8980650Z ##[endgroup] 2025-12-04T09:49:42.8999952Z /home/ec2-user/actions-runner/_work/_temp/5ae6f1ff-b779-4a96-8858-33f3a897218f.sh: line 1: kill: (52410) - No such process 2025-12-04T09:49:42.9016215Z ##[error]Process completed with exit code 1. 2025-12-04T09:49:42.9101865Z Prepare all required actions 2025-12-04T09:49:42.9102243Z Getting action download info 2025-12-04T09:49:43.0964488Z Download action repository 'seemethere/upload-artifact-s3@v5' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-12-04T09:49:43.3306558Z Download action repository 'actions/upload-artifact@v4' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-12-04T09:49:43.8472179Z ##[group]Run ./.github/actions/upload-test-artifacts 2025-12-04T09:49:43.8472404Z with: 2025-12-04T09:49:43.8472673Z file-suffix: test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344 2025-12-04T09:49:43.8472981Z s3-bucket: gha-artifacts 2025-12-04T09:49:43.8473148Z env: 2025-12-04T09:49:43.8473297Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:43.8473472Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:43.8473746Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:43.8474042Z DEVICE_NAME: 2025-12-04T09:49:43.8474237Z DEVICE_TYPE: 2025-12-04T09:49:43.8474383Z ##[endgroup] 2025-12-04T09:49:43.8490956Z ##[group]Run # Remove any previous test jsons if they exist 2025-12-04T09:49:43.8491248Z # Remove any previous test jsons if they exist 2025-12-04T09:49:43.8491493Z rm -f test-jsons-*.zip 2025-12-04T09:49:43.8491767Z zip -r "test-jsons-${FILE_SUFFIX}.zip" test/test-reports -i '*.json' 2025-12-04T09:49:43.8496371Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:43.8496597Z env: 2025-12-04T09:49:43.8496754Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:43.8496935Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:43.8497215Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:43.8497547Z DEVICE_NAME: 2025-12-04T09:49:43.8497712Z DEVICE_TYPE: 2025-12-04T09:49:43.8497982Z FILE_SUFFIX: test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344 2025-12-04T09:49:43.8498289Z ##[endgroup] 2025-12-04T09:49:43.8685076Z adding: test/test-reports/inference_huggingface.json (deflated 99%) 2025-12-04T09:49:43.8702226Z ##[group]Run # Remove any previous test reports if they exist 2025-12-04T09:49:43.8702524Z # Remove any previous test reports if they exist 2025-12-04T09:49:43.8702771Z rm -f test-reports-*.zip 2025-12-04T09:49:43.8703053Z zip -r "test-reports-${FILE_SUFFIX}.zip" test/test-reports -i '*.xml' -i '*.csv' 2025-12-04T09:49:43.8707223Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:43.8707464Z env: 2025-12-04T09:49:43.8707614Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:43.8707801Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:43.8708081Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:43.8708374Z DEVICE_NAME: 2025-12-04T09:49:43.8708519Z DEVICE_TYPE: 2025-12-04T09:49:43.8708788Z FILE_SUFFIX: test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344 2025-12-04T09:49:43.8709156Z ##[endgroup] 2025-12-04T09:49:43.8763286Z adding: test/test-reports/inference_huggingface.csv (deflated 62%) 2025-12-04T09:49:43.8763873Z adding: test/test-reports/inference_huggingface_graph_breaks.csv (deflated 85%) 2025-12-04T09:49:43.8764420Z adding: test/test-reports/inference_huggingface_graph_break_deduped.csv (deflated 65%) 2025-12-04T09:49:43.8784912Z ##[group]Run # Remove any previous usage logs if they exist 2025-12-04T09:49:43.8785265Z # Remove any previous usage logs if they exist 2025-12-04T09:49:43.8785497Z rm -f logs-*.zip 2025-12-04T09:49:43.8785721Z zip "logs-${FILE_SUFFIX}.zip" 'usage_log.txt' || true 2025-12-04T09:49:43.8786026Z zip -r "logs-${FILE_SUFFIX}.zip" test/test-reports -i '*.log' || true 2025-12-04T09:49:43.8789759Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:43.8789990Z env: 2025-12-04T09:49:43.8790141Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:43.8790315Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:43.8790698Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:43.8790997Z DEVICE_NAME: 2025-12-04T09:49:43.8791143Z DEVICE_TYPE: 2025-12-04T09:49:43.8791417Z FILE_SUFFIX: test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344 2025-12-04T09:49:43.8791712Z ##[endgroup] 2025-12-04T09:49:43.8834230Z adding: usage_log.txt (deflated 58%) 2025-12-04T09:49:43.8846559Z 2025-12-04T09:49:43.8848633Z zip error: Nothing to do! (logs-test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344.zip) 2025-12-04T09:49:43.8865033Z ##[group]Run # Remove any previous debugging artifacts if they exist 2025-12-04T09:49:43.8865373Z # Remove any previous debugging artifacts if they exist 2025-12-04T09:49:43.8865613Z rm -f debug-*.zip 2025-12-04T09:49:43.8865803Z if [ -d 'test/debug' ]; then 2025-12-04T09:49:43.8866029Z  zip -r "debug-${FILE_SUFFIX}.zip" test/debug 2025-12-04T09:49:43.8866381Z fi 2025-12-04T09:49:43.8869969Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:43.8870211Z env: 2025-12-04T09:49:43.8870369Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:43.8870541Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:43.8870829Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:43.8871131Z DEVICE_NAME: 2025-12-04T09:49:43.8871293Z DEVICE_TYPE: 2025-12-04T09:49:43.8871559Z FILE_SUFFIX: test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344 2025-12-04T09:49:43.8871864Z ##[endgroup] 2025-12-04T09:49:43.8937116Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T09:49:43.8937327Z with: 2025-12-04T09:49:43.8937486Z s3-bucket: gha-artifacts 2025-12-04T09:49:43.8937696Z s3-prefix: pytorch/pytorch/19923066595/1/artifact 2025-12-04T09:49:43.8937908Z retention-days: 14 2025-12-04T09:49:43.8938079Z if-no-files-found: warn 2025-12-04T09:49:43.8938272Z path: test-jsons-*.zip 2025-12-04T09:49:43.8938432Z name: artifact 2025-12-04T09:49:43.8938589Z region: us-east-1 2025-12-04T09:49:43.8938742Z env: 2025-12-04T09:49:43.8938884Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:43.8939059Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:43.8939337Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:43.8939630Z DEVICE_NAME: 2025-12-04T09:49:43.8939781Z DEVICE_TYPE: 2025-12-04T09:49:43.8939933Z ##[endgroup] 2025-12-04T09:49:44.1535831Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T09:49:44.1536280Z With the provided path, there will be 1 file uploaded 2025-12-04T09:49:44.1541531Z Uploading to s3 prefix: pytorch/pytorch/19923066595/1/artifact 2025-12-04T09:49:44.1565835Z Starting upload of test-jsons-test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344.zip 2025-12-04T09:49:44.2603124Z Finished upload of test-jsons-test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344.zip 2025-12-04T09:49:44.2751163Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T09:49:44.2751373Z with: 2025-12-04T09:49:44.2751538Z s3-bucket: gha-artifacts 2025-12-04T09:49:44.2751748Z s3-prefix: pytorch/pytorch/19923066595/1/artifact 2025-12-04T09:49:44.2751960Z retention-days: 14 2025-12-04T09:49:44.2752129Z if-no-files-found: error 2025-12-04T09:49:44.2752306Z path: test-reports-*.zip 2025-12-04T09:49:44.2752542Z name: artifact 2025-12-04T09:49:44.2752704Z region: us-east-1 2025-12-04T09:49:44.2752861Z env: 2025-12-04T09:49:44.2753002Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:44.2753178Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:44.2753462Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:44.2753758Z DEVICE_NAME: 2025-12-04T09:49:44.2753910Z DEVICE_TYPE: 2025-12-04T09:49:44.2754067Z ##[endgroup] 2025-12-04T09:49:44.5598546Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T09:49:44.5602060Z With the provided path, there will be 1 file uploaded 2025-12-04T09:49:44.5602446Z Uploading to s3 prefix: pytorch/pytorch/19923066595/1/artifact 2025-12-04T09:49:44.5630737Z Starting upload of test-reports-test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344.zip 2025-12-04T09:49:44.6633026Z Finished upload of test-reports-test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344.zip 2025-12-04T09:49:44.6781632Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T09:49:44.6781852Z with: 2025-12-04T09:49:44.6782011Z s3-bucket: gha-artifacts 2025-12-04T09:49:44.6782225Z s3-prefix: pytorch/pytorch/19923066595/1/artifact 2025-12-04T09:49:44.6782441Z retention-days: 14 2025-12-04T09:49:44.6782614Z if-no-files-found: ignore 2025-12-04T09:49:44.6782794Z path: logs-*.zip 2025-12-04T09:49:44.6782944Z name: artifact 2025-12-04T09:49:44.6783097Z region: us-east-1 2025-12-04T09:49:44.6783253Z env: 2025-12-04T09:49:44.6783406Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:44.6783598Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:44.6783882Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:44.6784176Z DEVICE_NAME: 2025-12-04T09:49:44.6784329Z DEVICE_TYPE: 2025-12-04T09:49:44.6784479Z ##[endgroup] 2025-12-04T09:49:44.9157140Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T09:49:44.9158706Z With the provided path, there will be 1 file uploaded 2025-12-04T09:49:44.9159026Z Uploading to s3 prefix: pytorch/pytorch/19923066595/1/artifact 2025-12-04T09:49:44.9186356Z Starting upload of logs-test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344.zip 2025-12-04T09:49:45.0177229Z Finished upload of logs-test-cpu_inductor_amp_freezing_huggingface-1-1-linux.8xlarge.amx_57118563344.zip 2025-12-04T09:49:45.0371984Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T09:49:45.0372199Z with: 2025-12-04T09:49:45.0372362Z s3-bucket: gha-artifacts 2025-12-04T09:49:45.0372601Z s3-prefix: pytorch/pytorch/19923066595/1/artifact 2025-12-04T09:49:45.0372809Z retention-days: 14 2025-12-04T09:49:45.0372976Z if-no-files-found: ignore 2025-12-04T09:49:45.0373152Z path: debug-*.zip 2025-12-04T09:49:45.0373297Z name: artifact 2025-12-04T09:49:45.0373448Z region: us-east-1 2025-12-04T09:49:45.0373599Z env: 2025-12-04T09:49:45.0373733Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:45.0373903Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:45.0374186Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:45.0374479Z DEVICE_NAME: 2025-12-04T09:49:45.0374622Z DEVICE_TYPE: 2025-12-04T09:49:45.0374770Z ##[endgroup] 2025-12-04T09:49:45.2754800Z No files were found with the provided path: debug-*.zip. No artifacts will be uploaded. 2025-12-04T09:49:45.2915880Z ##[group]Run # shellcheck disable=SC2156 2025-12-04T09:49:45.2916136Z # shellcheck disable=SC2156 2025-12-04T09:49:45.2916502Z find . -iname "core.[1-9]*" -exec docker exec "${DOCKER_CONTAINER_ID}" sh -c "gdb python {} -ex 'bt' -ex 'q'" \; 2025-12-04T09:49:45.2921249Z shell: /usr/bin/bash -e {0} 2025-12-04T09:49:45.2921435Z env: 2025-12-04T09:49:45.2921759Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:45.2921937Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:45.2922231Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:45.2922607Z DEVICE_NAME: 2025-12-04T09:49:45.2922764Z DEVICE_TYPE: 2025-12-04T09:49:45.2922927Z ##[endgroup] 2025-12-04T09:49:45.4669575Z Prepare all required actions 2025-12-04T09:49:45.4669915Z Getting action download info 2025-12-04T09:49:45.6384417Z Download action repository 'actions/setup-python@v6' (SHA:83679a892e2d95755f2dac6acb0bfd1e9ac5d548) 2025-12-04T09:49:46.0424876Z ##[group]Run ./.github/actions/upload-utilization-stats 2025-12-04T09:49:46.0425119Z with: 2025-12-04T09:49:46.0425274Z job_id: 57118563344 2025-12-04T09:49:46.0425632Z job_name: periodic-dynamo-benchmarks-cpu-test / test (cpu_inductor_amp_freezing_huggingface, 1, 1, linux.8xlarge.amx) 2025-12-04T09:49:46.0426028Z workflow_name: inductor-periodic 2025-12-04T09:49:46.0426218Z workflow_run_id: 19923066595 2025-12-04T09:49:46.0426396Z workflow_attempt: 1 2025-12-04T09:49:46.0426552Z env: 2025-12-04T09:49:46.0426692Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:46.0426868Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:46.0427149Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:46.0427479Z DEVICE_NAME: 2025-12-04T09:49:46.0427628Z DEVICE_TYPE: 2025-12-04T09:49:46.0427782Z ##[endgroup] 2025-12-04T09:49:46.0453610Z ##[group]Run actions/setup-python@v6 2025-12-04T09:49:46.0453811Z with: 2025-12-04T09:49:46.0453964Z python-version: 3.10 2025-12-04T09:49:46.0454125Z check-latest: false 2025-12-04T09:49:46.0454394Z token: *** 2025-12-04T09:49:46.0454554Z update-environment: true 2025-12-04T09:49:46.0454735Z allow-prereleases: false 2025-12-04T09:49:46.0454909Z freethreaded: false 2025-12-04T09:49:46.0455069Z env: 2025-12-04T09:49:46.0455215Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:46.0455377Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:46.0455653Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:46.0455951Z DEVICE_NAME: 2025-12-04T09:49:46.0456096Z DEVICE_TYPE: 2025-12-04T09:49:46.0456246Z ##[endgroup] 2025-12-04T09:49:46.1583718Z ##[group]Installed versions 2025-12-04T09:49:46.1598451Z Version 3.10 was not found in the local cache 2025-12-04T09:49:46.1714731Z (node:72502) [DEP0040] DeprecationWarning: The `punycode` module is deprecated. Please use a userland alternative instead. 2025-12-04T09:49:46.1715489Z (Use `node --trace-deprecation ...` to show where the warning was created) 2025-12-04T09:49:46.5278476Z ##[error]The version '3.10' with architecture 'x64' was not found for this operating system. The list of all available versions can be found here: https://raw.githubusercontent.com/actions/python-versions/main/versions-manifest.json 2025-12-04T09:49:46.5403891Z ##[group]Run pytorch/test-infra/.github/actions/teardown-linux@main 2025-12-04T09:49:46.5404181Z with: 2025-12-04T09:49:46.5404342Z env: 2025-12-04T09:49:46.5404500Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:46.5404697Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:46.5405015Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:46.5405314Z DEVICE_NAME: 2025-12-04T09:49:46.5405464Z DEVICE_TYPE: 2025-12-04T09:49:46.5405685Z ##[endgroup] 2025-12-04T09:49:46.5416115Z ##[group]Run set -eou pipefail 2025-12-04T09:49:46.5416326Z set -eou pipefail 2025-12-04T09:49:46.5416496Z  2025-12-04T09:49:46.5416724Z echo "Holding runner for 2 hours until all ssh sessions have logged out" 2025-12-04T09:49:46.5416993Z for _ in $(seq 1440); do 2025-12-04T09:49:46.5417199Z  # Break if no ssh session exists anymore 2025-12-04T09:49:46.5417475Z  if [ "$(who)" = "" ]; then 2025-12-04T09:49:46.5417688Z  break 2025-12-04T09:49:46.5417840Z  fi 2025-12-04T09:49:46.5417990Z  echo "." 2025-12-04T09:49:46.5418149Z  sleep 5 2025-12-04T09:49:46.5418293Z done 2025-12-04T09:49:46.5423370Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:46.5423603Z env: 2025-12-04T09:49:46.5423820Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:46.5424001Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:46.5424284Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:46.5424567Z DEVICE_NAME: 2025-12-04T09:49:46.5424724Z DEVICE_TYPE: 2025-12-04T09:49:46.5424878Z ##[endgroup] 2025-12-04T09:49:46.5446765Z Holding runner for 2 hours until all ssh sessions have logged out 2025-12-04T09:49:46.5567192Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T09:49:46.5567526Z # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T09:49:46.5567805Z # shellcheck disable=SC2046 2025-12-04T09:49:46.5568022Z docker stop $(docker ps -q) || true 2025-12-04T09:49:46.5568235Z # Prune all of the docker images 2025-12-04T09:49:46.5568428Z docker system prune -af 2025-12-04T09:49:46.5572234Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:49:46.5572476Z env: 2025-12-04T09:49:46.5572625Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:49:46.5572805Z HAS_NVIDIA_GPU: false 2025-12-04T09:49:46.5573091Z DOCKER_CONTAINER_ID: 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:46.5573383Z DEVICE_NAME: 2025-12-04T09:49:46.5573538Z DEVICE_TYPE: 2025-12-04T09:49:46.5573688Z ##[endgroup] 2025-12-04T09:49:57.2670663Z 25f649d00d6e 2025-12-04T09:49:57.5645617Z Deleted Containers: 2025-12-04T09:49:57.5649731Z 25f649d00d6e23eba86570d7ba13b6f3904fca5c84b63b5f7634f6b5bb236359 2025-12-04T09:49:57.5653635Z 2025-12-04T09:50:04.4043648Z Deleted Images: 2025-12-04T09:50:04.4044394Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:50:04.4045167Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image@sha256:b178ee928adfcab963ebbc6ea05ea1a0f0c605bb095e9076f010bf92d150869d 2025-12-04T09:50:04.4045706Z deleted: sha256:40bd34edc3497e05e2150804a15313235b6426bcc4b8c19c03c3daf5524be913 2025-12-04T09:50:04.4046073Z deleted: sha256:947006dce32e090f6b67d556f519398b957121651577f10819c7b179a5b6b938 2025-12-04T09:50:04.4046427Z deleted: sha256:dca16630fc5508e008012b760ac178fac5af74cb679706a3d9898100501d6f8e 2025-12-04T09:50:04.4046821Z deleted: sha256:a35e2e4cfaaa05e1f193fbe67c4151ded52737cb30c75429050cb7cd7777da23 2025-12-04T09:50:04.4047199Z deleted: sha256:d91e1a0a76e64abd7933fc58b8a553d76ac6cd82269c1bf701ade4e2c5ab58ff 2025-12-04T09:50:04.4047602Z deleted: sha256:f96fff40164ffa27ae535a4e5da2bbf2b18ec293366adc73519c830606f07baa 2025-12-04T09:50:04.4048347Z deleted: sha256:649089a3156b98c2cbf16b0af2b1ebdaa8644f9bf90534d9a1f7657b78b9f90e 2025-12-04T09:50:04.4048791Z deleted: sha256:aebabb7ea7535ad2aca44b23eb308b16d5de44f5b412574f94620d923436f42e 2025-12-04T09:50:04.4049161Z deleted: sha256:bcb08dc52ae55732fc33ab8c011347049a9279cc5f45a50d960c95d3a7cd5a3d 2025-12-04T09:50:04.4049533Z deleted: sha256:aa538cda5b22ea6e81fb69326637459a1356cce39b618657a075ce25d9124876 2025-12-04T09:50:04.4049911Z deleted: sha256:e28671ddaf02cbac7c1dcfdb2b08a40d3b7e2dab69c9a9e6c48ede3d8d550d66 2025-12-04T09:50:04.4050275Z deleted: sha256:ba797a4f759946d2fa7414073a5590cc12337aaa43197c35326ce456a3f84adf 2025-12-04T09:50:04.4050630Z deleted: sha256:25b28b66e2a53c3498a56aff9f543bf3a4641534c9e69741c673e9d018c652d8 2025-12-04T09:50:04.4050995Z deleted: sha256:860240198a7e7dcec697a58bcf979cfde64334033f2e97d4c9609277e051558c 2025-12-04T09:50:04.4051359Z deleted: sha256:39de54a9c331b8da952419605b26c35862b4cbb8c8a0967cc536a02b96eda874 2025-12-04T09:50:04.4051784Z deleted: sha256:6f78957316e0887707bc6534ea2aa43d1cd016130ec6a0d3148370b879696e1a 2025-12-04T09:50:04.4052152Z deleted: sha256:8656f7aa172ef2a37ec1524ac6f3ecd2b3fd9675a78a777b9a902d0f26047ea9 2025-12-04T09:50:04.4052531Z deleted: sha256:255344ff6c764e507ea2f21fbbdc6e0a1aaab0e5e0835935494c9df7d620e687 2025-12-04T09:50:04.4052947Z deleted: sha256:b6a1addb4370e1500d2fd65255b80922ad308c95a846a3460be730442d75fbe4 2025-12-04T09:50:04.4053376Z deleted: sha256:8dafa1eff3c9aada8b6725626ba0dd7de62719fb5ff307be3005cf2c6e72e2df 2025-12-04T09:50:04.4053755Z deleted: sha256:9e3fde92e8ff4f05cadac548766ef5ef9e8beb3c89cb48b4e2a17843ff60d649 2025-12-04T09:50:04.4054124Z deleted: sha256:7169a6f70d1f624733e52107b4913effe35e1e63c99388164b0afd2345658178 2025-12-04T09:50:04.4054480Z deleted: sha256:4c0d4c67766b690c1f1834310cd3898a900cd9742b7157de7e8f64d651600e35 2025-12-04T09:50:04.4054830Z deleted: sha256:7343a9c5c68976b0328278e691ddbcf26f619f989ca2a51933ac8bf637d1d69c 2025-12-04T09:50:04.4055201Z deleted: sha256:b790f29be6db3408a1a53fabdc69e76bb38cef221474a4f3aef1e233bb85d6d3 2025-12-04T09:50:04.4055565Z deleted: sha256:ed0831c232a6a410214e113695ab9288063623b523a6032a05b443fa3e9a68ae 2025-12-04T09:50:04.4055913Z deleted: sha256:554ed49d988a9f7a1381d193479542175c40d702a22d246ecef7958a8662a2eb 2025-12-04T09:50:04.4056284Z deleted: sha256:47b04d260bfb2b1f68a2f12aa70926ae2a2a53be6f4d8811d6fbe17fd7f13e17 2025-12-04T09:50:04.4056656Z deleted: sha256:6eb59e87109db04da60df057ff6b232045b6df4e2729ce7534a7a77fbe663761 2025-12-04T09:50:04.4057026Z deleted: sha256:0d9763e3d6e83eee92752c3a5dcfce74703dfa0aba42042761d7600edd95baa5 2025-12-04T09:50:04.4057383Z deleted: sha256:5811882cf3f14a50f47dc91a87d1be09daf3c20494aaf7e7343554066e9f90b7 2025-12-04T09:50:04.4057740Z deleted: sha256:4a6f4f545e9062949802a6c597609ef772633eae7d6925d41823d2b1a8991d4d 2025-12-04T09:50:04.4058141Z deleted: sha256:69a716cb9dc95d7d58d1416f4d127e49885c5d96882dc593686a6eb97200582a 2025-12-04T09:50:04.4058525Z deleted: sha256:0542610effe51f46308d4570438af559e4f6a263f38714ea04c78e4661d1909a 2025-12-04T09:50:04.4058879Z deleted: sha256:86dcdb716a38a00a400b483f42e4ac58714a70e89f5671b50ca5c45d8f577217 2025-12-04T09:50:04.4059254Z deleted: sha256:a9aaf6de8156c1b1a8ece816adbc7d2ed6edabb95e36732737ca293b26619b40 2025-12-04T09:50:04.4059633Z deleted: sha256:57ce7fcbaaa86795f8f7e6290c0aaa8622e13c9c82191d34952c0a4d2e203326 2025-12-04T09:50:04.4059996Z deleted: sha256:a90408b5f4c06a3ba881facc34fa55beb4ae4e51823e7f8d933e611f68465f4a 2025-12-04T09:50:04.4060370Z deleted: sha256:1ac1aab20bc98902f03ee9c657ff435af431f40a81ac9cebf1cd91d9041bc6b1 2025-12-04T09:50:04.4060743Z deleted: sha256:89e6c1c76b21fbe98c153a32627fa1b2337a8caa744b09b28253b67cb84012cb 2025-12-04T09:50:04.4061107Z deleted: sha256:3765d58e17aa81539a8e1eb3cb45914b48f8fb53f7ac12c1bf06c9390e274a93 2025-12-04T09:50:04.4061457Z deleted: sha256:6056a51594e8137d3425100cc233c8ab0f09dded2dc491ba083e8d7c0e1f1143 2025-12-04T09:50:04.4061817Z deleted: sha256:43a55512a68fc9eb9c622a73682dbf9f3c83f1b499287d9e12151f559edabe85 2025-12-04T09:50:04.4062261Z deleted: sha256:a5ad8e88a7e08768a61feeb801fb5a4c5041587feb2a2c77c35b285e869c25c9 2025-12-04T09:50:04.4062622Z deleted: sha256:55fd389078a5792f567a3592821e38fcfa0b0e114fa43248c4528c4d85cf6459 2025-12-04T09:50:04.4062983Z deleted: sha256:144567b58c97f1b5a75d54c3a1514a26ad5ed2dc7988e166316e324a76e76aa6 2025-12-04T09:50:04.4063335Z deleted: sha256:049fbb3123a92952eba68198938b7b29203fcb3d74dfc74e67f9929999576f19 2025-12-04T09:50:04.4063690Z deleted: sha256:22607d21e148d19a45f3f7afe927a15041b9a5663430f5580cc993c4bf9dd7dc 2025-12-04T09:50:04.4064027Z deleted: sha256:63a29311926064245f6d4a4661513586c8179b1cb460ceb91184c0d1114a742d 2025-12-04T09:50:04.4064383Z deleted: sha256:0072101bc0e6bbe2457f4a2d0de953ae643fc2aa6adc69bb59da9cbf6006c8a9 2025-12-04T09:50:04.4064751Z deleted: sha256:35c8f84104fcf39bc1be4682f46dc8b04cb1c5e4c26f66aac039cb42a3f08128 2025-12-04T09:50:04.4065120Z deleted: sha256:c6da65c085b8e72b2e240a3a7d79c4aa166c247737f5916587e51d40acff1ab8 2025-12-04T09:50:04.4065507Z deleted: sha256:b8b5ced5e8550748e26855aad0bb30864c4c435c64694097e63a9a686ef90ba0 2025-12-04T09:50:04.4065864Z deleted: sha256:9905f824759cfae96d29353305bc8a6130a7316db29dfc3efeb8ba9c7cbb5172 2025-12-04T09:50:04.4066220Z deleted: sha256:2b93a73a0fc57cdcfab2a04a30c9b313d84e5e447d076575212f198508242a00 2025-12-04T09:50:04.4066573Z deleted: sha256:baa73842b47530e5d1421010d895ed9059a567516a7e85b10943a6b5fe6fbe41 2025-12-04T09:50:04.4066948Z deleted: sha256:db79627a360d2322d58fb57c98268676011788961d4682ba034f17c9f5fdd855 2025-12-04T09:50:04.4067306Z deleted: sha256:8439f69d5eab9d42b15df43b938f56c690b18cd5e3884f4b80ddc9edf9e43822 2025-12-04T09:50:04.4067684Z deleted: sha256:e95ccdb10d43bbb6f2d6416bca8e5ad7d76a39985431aa5e40ff192ebeb52883 2025-12-04T09:50:04.4068036Z deleted: sha256:d878081c591fd25e8e494f4565e67f341cb3c459b21596ad8e8a707ff9ca63e6 2025-12-04T09:50:04.4068386Z deleted: sha256:2d594947cf73e58a20e36d52acf22f3d31b7a43291d7017c1ac71ea38c4aa9d1 2025-12-04T09:50:04.4068741Z deleted: sha256:933d315b5f20ac941653d01f2b896cc7710564c9b9c4bf9fd8e850115180c6f9 2025-12-04T09:50:04.4069100Z deleted: sha256:0edebbabed274b9e777a5933aaa21ed58569a412c97423c365eff4645e773981 2025-12-04T09:50:04.4069478Z deleted: sha256:744e1cebe181659b0a93ac857d4c16af4c2ce29f3227368eacd7ef18b69e06a7 2025-12-04T09:50:04.4069841Z deleted: sha256:71393f74ab60b183866b9f5995f85aaa65613968c611c8dc15d043befbac662b 2025-12-04T09:50:04.4070205Z deleted: sha256:741a07a29d03bdecf1fb4dfc8f1654a136064f14e5ded89a1497f476ff24623d 2025-12-04T09:50:04.4070569Z deleted: sha256:83df0b5b82e0926d6046012c3b5098dae9aa37177323a33d6eda09e392844b86 2025-12-04T09:50:04.4070953Z deleted: sha256:65b3d6509a37164c67b36965f4ca77ccb1a8e62bdd8f058e2d0886c82ba0940f 2025-12-04T09:50:04.4071332Z deleted: sha256:a38d6cb5dacc51c3e614f4ed42c05cdc53fc489f6e3dafc3b7d0eb39d3250908 2025-12-04T09:50:04.4071706Z deleted: sha256:c3de4f46ac5273a8218c6c5325247b0bd17f12037cfb49f5101a2b5085746704 2025-12-04T09:50:04.4072065Z deleted: sha256:7f3448d255710b28f79767ded93dcf11e0c1a941c7579d8752f4f0bc589aa0b5 2025-12-04T09:50:04.4072434Z deleted: sha256:fd33fe89098b4a8897870a9a316cd91868ab19f42f800d5b512e679356791e62 2025-12-04T09:50:04.4072803Z deleted: sha256:df2c12f62f5ceb2670bf7cb8aa53f268d11dbbfbf25f4b4d567673702cc3fc30 2025-12-04T09:50:04.4073168Z deleted: sha256:a3d7900a6180e4612708429f9c0edba3e999bb43fd67fd3406ae088ea7016a47 2025-12-04T09:50:04.4073535Z deleted: sha256:f592639a1c1bf8da30a1c21866ee22e76b33ae029ee236c7fd20b1ecc4716775 2025-12-04T09:50:04.4073911Z deleted: sha256:e11e4a9756b2bd4a6e62cde476dc36c6528d71a74e8e7570b271784bc494ab40 2025-12-04T09:50:04.4074281Z deleted: sha256:4c03ab513fc4cf698dd61fc37312e57abf7c4e44d4d4ddab126f38804fddec75 2025-12-04T09:50:04.4074649Z deleted: sha256:b4d3d8b5232e0bac87fa78b9c77997cbe8798faffce2aac259b820ab387ce4ad 2025-12-04T09:50:04.4075040Z deleted: sha256:dfda7f6e86c7217a3eb1fe089352248b5d84ca4f62734abce70beb4a6fc90779 2025-12-04T09:50:04.4075400Z deleted: sha256:00f41b846c66bc26a6381fe3663c611a2f539ae3b7e46e007436afd3335a7517 2025-12-04T09:50:04.4075788Z deleted: sha256:0f35c2638249a4525b29668586b079b4ee01f73bbe33732c9da7cfa7b6afb480 2025-12-04T09:50:04.4076133Z deleted: sha256:7441da072ac3269ddb67441b265fdf82904871335b5c19bb381d3622a95b7745 2025-12-04T09:50:04.4076483Z deleted: sha256:ba70851a3aedb994e8d7f8657f8f108bd90a19272843f389e6a8af48a61db720 2025-12-04T09:50:04.4076842Z deleted: sha256:1984ce6f5cabbd8c86386a8d9b5365923b92bc7ccb9aace9b4fe9570213bbbe1 2025-12-04T09:50:04.4077188Z deleted: sha256:73974f74b436f39a2fdb6461b1e3f7c3e41c73325776fa71d16b942a5b4a365b 2025-12-04T09:50:04.4077408Z 2025-12-04T09:50:04.4077492Z Total reclaimed space: 53.28GB 2025-12-04T09:50:04.4149253Z Post job cleanup. 2025-12-04T09:50:04.4174074Z Post job cleanup. 2025-12-04T09:50:04.5234088Z (node:72651) [DEP0040] DeprecationWarning: The `punycode` module is deprecated. Please use a userland alternative instead. 2025-12-04T09:50:04.5235788Z (Use `node --trace-deprecation ...` to show where the warning was created) 2025-12-04T09:50:04.5351586Z Post job cleanup. 2025-12-04T09:50:04.5379917Z Post job cleanup. 2025-12-04T09:50:04.6199492Z [command]/usr/bin/git version 2025-12-04T09:50:04.6237688Z git version 2.50.1 2025-12-04T09:50:04.6271989Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/d4d1cef2-cf05-4f56-82d6-4d11fabfaa7b/.gitconfig' 2025-12-04T09:50:04.6298583Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/d4d1cef2-cf05-4f56-82d6-4d11fabfaa7b' before making global git config changes 2025-12-04T09:50:04.6304819Z Adding repository directory to the temporary git global config as a safe directory 2025-12-04T09:50:04.6310440Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:50:04.6357122Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-12-04T09:50:04.6398743Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-12-04T09:50:04.6707369Z Entering 'android/libs/fbjni' 2025-12-04T09:50:04.6759149Z Entering 'third_party/FP16' 2025-12-04T09:50:04.6813835Z Entering 'third_party/FXdiv' 2025-12-04T09:50:04.6867483Z Entering 'third_party/NNPACK' 2025-12-04T09:50:04.6920973Z Entering 'third_party/NVTX' 2025-12-04T09:50:04.6980315Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:50:04.7030740Z Entering 'third_party/XNNPACK' 2025-12-04T09:50:04.7099392Z Entering 'third_party/aiter' 2025-12-04T09:50:04.7154182Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:50:04.7217574Z Entering 'third_party/benchmark' 2025-12-04T09:50:04.7271314Z Entering 'third_party/composable_kernel' 2025-12-04T09:50:04.7333190Z Entering 'third_party/cpp-httplib' 2025-12-04T09:50:04.7384152Z Entering 'third_party/cpuinfo' 2025-12-04T09:50:04.7441864Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:50:04.7492612Z Entering 'third_party/cutlass' 2025-12-04T09:50:04.7555127Z Entering 'third_party/fbgemm' 2025-12-04T09:50:04.7604930Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:50:04.7661991Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:50:04.7722202Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:50:04.7774629Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:50:04.7836885Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:50:04.7888136Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:50:04.7936341Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:50:04.7991868Z Entering 'third_party/flash-attention' 2025-12-04T09:50:04.8044923Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:50:04.8102845Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:50:04.8161952Z Entering 'third_party/flatbuffers' 2025-12-04T09:50:04.8218875Z Entering 'third_party/fmt' 2025-12-04T09:50:04.8269196Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:50:04.8323921Z Entering 'third_party/gloo' 2025-12-04T09:50:04.8377035Z Entering 'third_party/googletest' 2025-12-04T09:50:04.8430958Z Entering 'third_party/ideep' 2025-12-04T09:50:04.8482159Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:50:04.8540682Z Entering 'third_party/ittapi' 2025-12-04T09:50:04.8596159Z Entering 'third_party/kineto' 2025-12-04T09:50:04.8648079Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:50:04.8700768Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:50:04.8754334Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:50:04.8806941Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:50:04.8858704Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:50:04.8912830Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:50:04.8970290Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:50:04.9027966Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:50:04.9077984Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:50:04.9129772Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:50:04.9184257Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:50:04.9235288Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:50:04.9297158Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:50:04.9351609Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:50:04.9403304Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:50:04.9458310Z Entering 'third_party/kleidiai' 2025-12-04T09:50:04.9511662Z Entering 'third_party/mimalloc' 2025-12-04T09:50:04.9567051Z Entering 'third_party/nlohmann' 2025-12-04T09:50:04.9621776Z Entering 'third_party/onnx' 2025-12-04T09:50:04.9689293Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:50:04.9746256Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:50:04.9800101Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:50:04.9855463Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:50:04.9912984Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:50:04.9966956Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:50:05.0026379Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:50:05.0080934Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:50:05.0130593Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:50:05.0179805Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:50:05.0233473Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:50:05.0290757Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:50:05.0358457Z Entering 'third_party/pocketfft' 2025-12-04T09:50:05.0412305Z Entering 'third_party/protobuf' 2025-12-04T09:50:05.0465953Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:50:05.0519258Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:50:05.0578889Z Entering 'third_party/psimd' 2025-12-04T09:50:05.0633139Z Entering 'third_party/pthreadpool' 2025-12-04T09:50:05.0687535Z Entering 'third_party/pybind11' 2025-12-04T09:50:05.0741022Z Entering 'third_party/python-peachpy' 2025-12-04T09:50:05.0795301Z Entering 'third_party/sleef' 2025-12-04T09:50:05.0848045Z Entering 'third_party/tensorpipe' 2025-12-04T09:50:05.0899218Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:50:05.0954447Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:50:05.1003638Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:50:05.1057632Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:50:05.1106583Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:50:05.1191094Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-12-04T09:50:05.1216120Z http.https://github.com/.extraheader 2025-12-04T09:50:05.1229005Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-12-04T09:50:05.1263404Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-12-04T09:50:05.1560831Z Entering 'android/libs/fbjni' 2025-12-04T09:50:05.1597026Z http.https://github.com/.extraheader 2025-12-04T09:50:05.1626485Z Entering 'third_party/FP16' 2025-12-04T09:50:05.1663740Z http.https://github.com/.extraheader 2025-12-04T09:50:05.1699557Z Entering 'third_party/FXdiv' 2025-12-04T09:50:05.1732457Z http.https://github.com/.extraheader 2025-12-04T09:50:05.1768560Z Entering 'third_party/NNPACK' 2025-12-04T09:50:05.1802320Z http.https://github.com/.extraheader 2025-12-04T09:50:05.1837734Z Entering 'third_party/NVTX' 2025-12-04T09:50:05.1870836Z http.https://github.com/.extraheader 2025-12-04T09:50:05.1911776Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:50:05.1942689Z http.https://github.com/.extraheader 2025-12-04T09:50:05.1980934Z Entering 'third_party/XNNPACK' 2025-12-04T09:50:05.2014031Z http.https://github.com/.extraheader 2025-12-04T09:50:05.2061232Z Entering 'third_party/aiter' 2025-12-04T09:50:05.2095663Z http.https://github.com/.extraheader 2025-12-04T09:50:05.2125770Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:50:05.2162151Z http.https://github.com/.extraheader 2025-12-04T09:50:05.2206708Z Entering 'third_party/benchmark' 2025-12-04T09:50:05.2238841Z http.https://github.com/.extraheader 2025-12-04T09:50:05.2273877Z Entering 'third_party/composable_kernel' 2025-12-04T09:50:05.2309100Z http.https://github.com/.extraheader 2025-12-04T09:50:05.2348706Z Entering 'third_party/cpp-httplib' 2025-12-04T09:50:05.2383430Z http.https://github.com/.extraheader 2025-12-04T09:50:05.2420739Z Entering 'third_party/cpuinfo' 2025-12-04T09:50:05.2456399Z http.https://github.com/.extraheader 2025-12-04T09:50:05.2495472Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:50:05.2525624Z http.https://github.com/.extraheader 2025-12-04T09:50:05.2560211Z Entering 'third_party/cutlass' 2025-12-04T09:50:05.2594523Z http.https://github.com/.extraheader 2025-12-04T09:50:05.2639406Z Entering 'third_party/fbgemm' 2025-12-04T09:50:05.2671210Z http.https://github.com/.extraheader 2025-12-04T09:50:05.2707124Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:50:05.2742417Z http.https://github.com/.extraheader 2025-12-04T09:50:05.2777992Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:50:05.2813693Z http.https://github.com/.extraheader 2025-12-04T09:50:05.2852236Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:50:05.2883994Z http.https://github.com/.extraheader 2025-12-04T09:50:05.2920991Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:50:05.2952483Z http.https://github.com/.extraheader 2025-12-04T09:50:05.2992823Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:50:05.3025955Z http.https://github.com/.extraheader 2025-12-04T09:50:05.3060202Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:50:05.3094372Z http.https://github.com/.extraheader 2025-12-04T09:50:05.3128586Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:50:05.3161300Z http.https://github.com/.extraheader 2025-12-04T09:50:05.3197919Z Entering 'third_party/flash-attention' 2025-12-04T09:50:05.3231850Z http.https://github.com/.extraheader 2025-12-04T09:50:05.3265497Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:50:05.3300034Z http.https://github.com/.extraheader 2025-12-04T09:50:05.3336608Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:50:05.3368130Z http.https://github.com/.extraheader 2025-12-04T09:50:05.3413076Z Entering 'third_party/flatbuffers' 2025-12-04T09:50:05.3447627Z http.https://github.com/.extraheader 2025-12-04T09:50:05.3482586Z Entering 'third_party/fmt' 2025-12-04T09:50:05.3515617Z http.https://github.com/.extraheader 2025-12-04T09:50:05.3545779Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:50:05.3580261Z http.https://github.com/.extraheader 2025-12-04T09:50:05.3622169Z Entering 'third_party/gloo' 2025-12-04T09:50:05.3660074Z http.https://github.com/.extraheader 2025-12-04T09:50:05.3691732Z Entering 'third_party/googletest' 2025-12-04T09:50:05.3726019Z http.https://github.com/.extraheader 2025-12-04T09:50:05.3762238Z Entering 'third_party/ideep' 2025-12-04T09:50:05.3797644Z http.https://github.com/.extraheader 2025-12-04T09:50:05.3826140Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:50:05.3864676Z http.https://github.com/.extraheader 2025-12-04T09:50:05.3906394Z Entering 'third_party/ittapi' 2025-12-04T09:50:05.3940056Z http.https://github.com/.extraheader 2025-12-04T09:50:05.3970308Z Entering 'third_party/kineto' 2025-12-04T09:50:05.4002019Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4035681Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:50:05.4070386Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4103413Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:50:05.4137486Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4172021Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:50:05.4208847Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4242061Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:50:05.4272368Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4310571Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:50:05.4344125Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4375152Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:50:05.4403500Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4442802Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:50:05.4477133Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4512855Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:50:05.4544581Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4577110Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:50:05.4611170Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4644781Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:50:05.4679948Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4711147Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:50:05.4743203Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4774955Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:50:05.4805666Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4849118Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:50:05.4878485Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4920532Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:50:05.4952469Z http.https://github.com/.extraheader 2025-12-04T09:50:05.4989177Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:50:05.5022983Z http.https://github.com/.extraheader 2025-12-04T09:50:05.5059778Z Entering 'third_party/kleidiai' 2025-12-04T09:50:05.5095708Z http.https://github.com/.extraheader 2025-12-04T09:50:05.5127565Z Entering 'third_party/mimalloc' 2025-12-04T09:50:05.5161733Z http.https://github.com/.extraheader 2025-12-04T09:50:05.5200063Z Entering 'third_party/nlohmann' 2025-12-04T09:50:05.5235775Z http.https://github.com/.extraheader 2025-12-04T09:50:05.5269782Z Entering 'third_party/onnx' 2025-12-04T09:50:05.5302028Z http.https://github.com/.extraheader 2025-12-04T09:50:05.5353788Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:50:05.5383874Z http.https://github.com/.extraheader 2025-12-04T09:50:05.5422655Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:50:05.5461369Z http.https://github.com/.extraheader 2025-12-04T09:50:05.5499567Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:50:05.5527928Z http.https://github.com/.extraheader 2025-12-04T09:50:05.5569120Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:50:05.5601169Z http.https://github.com/.extraheader 2025-12-04T09:50:05.5640700Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:50:05.5675566Z http.https://github.com/.extraheader 2025-12-04T09:50:05.5711857Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:50:05.5745508Z http.https://github.com/.extraheader 2025-12-04T09:50:05.5781283Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:50:05.5814124Z http.https://github.com/.extraheader 2025-12-04T09:50:05.5848344Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:50:05.5880009Z http.https://github.com/.extraheader 2025-12-04T09:50:05.5912973Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:50:05.5948489Z http.https://github.com/.extraheader 2025-12-04T09:50:05.5982565Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:50:05.6013349Z http.https://github.com/.extraheader 2025-12-04T09:50:05.6054400Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:50:05.6091145Z http.https://github.com/.extraheader 2025-12-04T09:50:05.6127366Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:50:05.6162450Z http.https://github.com/.extraheader 2025-12-04T09:50:05.6212578Z Entering 'third_party/pocketfft' 2025-12-04T09:50:05.6246087Z http.https://github.com/.extraheader 2025-12-04T09:50:05.6284598Z Entering 'third_party/protobuf' 2025-12-04T09:50:05.6315095Z http.https://github.com/.extraheader 2025-12-04T09:50:05.6354778Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:50:05.6385583Z http.https://github.com/.extraheader 2025-12-04T09:50:05.6422642Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:50:05.6458616Z http.https://github.com/.extraheader 2025-12-04T09:50:05.6497319Z Entering 'third_party/psimd' 2025-12-04T09:50:05.6528340Z http.https://github.com/.extraheader 2025-12-04T09:50:05.6563474Z Entering 'third_party/pthreadpool' 2025-12-04T09:50:05.6595941Z http.https://github.com/.extraheader 2025-12-04T09:50:05.6634209Z Entering 'third_party/pybind11' 2025-12-04T09:50:05.6666656Z http.https://github.com/.extraheader 2025-12-04T09:50:05.6703359Z Entering 'third_party/python-peachpy' 2025-12-04T09:50:05.6734465Z http.https://github.com/.extraheader 2025-12-04T09:50:05.6772180Z Entering 'third_party/sleef' 2025-12-04T09:50:05.6804638Z http.https://github.com/.extraheader 2025-12-04T09:50:05.6846259Z Entering 'third_party/tensorpipe' 2025-12-04T09:50:05.6877205Z http.https://github.com/.extraheader 2025-12-04T09:50:05.6909245Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:50:05.6942210Z http.https://github.com/.extraheader 2025-12-04T09:50:05.6976994Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:50:05.7010670Z http.https://github.com/.extraheader 2025-12-04T09:50:05.7049807Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:50:05.7080043Z http.https://github.com/.extraheader 2025-12-04T09:50:05.7117092Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:50:05.7149645Z http.https://github.com/.extraheader 2025-12-04T09:50:05.7185576Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:50:05.7217295Z http.https://github.com/.extraheader 2025-12-04T09:50:05.7279857Z [command]/usr/bin/git config --local --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:05.7328129Z [command]/usr/bin/git submodule foreach --recursive git config --local --show-origin --name-only --get-regexp remote.origin.url 2025-12-04T09:50:05.7643564Z Entering 'android/libs/fbjni' 2025-12-04T09:50:05.7664851Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T09:50:05.7679859Z Entering 'third_party/FP16' 2025-12-04T09:50:05.7702645Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T09:50:05.7720275Z Entering 'third_party/FXdiv' 2025-12-04T09:50:05.7745886Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T09:50:05.7763740Z Entering 'third_party/NNPACK' 2025-12-04T09:50:05.7785172Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T09:50:05.7800956Z Entering 'third_party/NVTX' 2025-12-04T09:50:05.7822243Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T09:50:05.7842485Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:50:05.7865780Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T09:50:05.7881199Z Entering 'third_party/XNNPACK' 2025-12-04T09:50:05.7902069Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T09:50:05.7928954Z Entering 'third_party/aiter' 2025-12-04T09:50:05.7950798Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T09:50:05.7968633Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:50:05.7988879Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T09:50:05.8016318Z Entering 'third_party/benchmark' 2025-12-04T09:50:05.8038866Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:50:05.8057629Z Entering 'third_party/composable_kernel' 2025-12-04T09:50:05.8080618Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T09:50:05.8106373Z Entering 'third_party/cpp-httplib' 2025-12-04T09:50:05.8127043Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T09:50:05.8147781Z Entering 'third_party/cpuinfo' 2025-12-04T09:50:05.8167701Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T09:50:05.8188363Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:50:05.8206786Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T09:50:05.8225406Z Entering 'third_party/cutlass' 2025-12-04T09:50:05.8246440Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T09:50:05.8277382Z Entering 'third_party/fbgemm' 2025-12-04T09:50:05.8297779Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T09:50:05.8318605Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:50:05.8340579Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T09:50:05.8357457Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:50:05.8377435Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T09:50:05.8402306Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:50:05.8421974Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T09:50:05.8441654Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:50:05.8463894Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T09:50:05.8487465Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:50:05.8507952Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T09:50:05.8525018Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:50:05.8548624Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T09:50:05.8563537Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:50:05.8587004Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T09:50:05.8609613Z Entering 'third_party/flash-attention' 2025-12-04T09:50:05.8633177Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T09:50:05.8652226Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:50:05.8672886Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T09:50:05.8693033Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:50:05.8712584Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T09:50:05.8740167Z Entering 'third_party/flatbuffers' 2025-12-04T09:50:05.8759841Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T09:50:05.8781373Z Entering 'third_party/fmt' 2025-12-04T09:50:05.8800267Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T09:50:05.8819477Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:50:05.8842867Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T09:50:05.8862135Z Entering 'third_party/gloo' 2025-12-04T09:50:05.8880619Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T09:50:05.8900150Z Entering 'third_party/googletest' 2025-12-04T09:50:05.8920211Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:50:05.8939964Z Entering 'third_party/ideep' 2025-12-04T09:50:05.8962000Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T09:50:05.8979779Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:50:05.8999869Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T09:50:05.9023973Z Entering 'third_party/ittapi' 2025-12-04T09:50:05.9049221Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T09:50:05.9066899Z Entering 'third_party/kineto' 2025-12-04T09:50:05.9090532Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T09:50:05.9107172Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:50:05.9126932Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T09:50:05.9144769Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:50:05.9168201Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T09:50:05.9188382Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:50:05.9207958Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T09:50:05.9225566Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:50:05.9245561Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T09:50:05.9262905Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:50:05.9283367Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T09:50:05.9300252Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:50:05.9322319Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T09:50:05.9339786Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:50:05.9361859Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T09:50:05.9379830Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:50:05.9401148Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:50:05.9419932Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:50:05.9443576Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T09:50:05.9459912Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:50:05.9480986Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T09:50:05.9497896Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:50:05.9520019Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T09:50:05.9539211Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:50:05.9558645Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T09:50:05.9576388Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:50:05.9598297Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T09:50:05.9626006Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:50:05.9649208Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T09:50:05.9664496Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:50:05.9686772Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T09:50:05.9707352Z Entering 'third_party/kleidiai' 2025-12-04T09:50:05.9728577Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T09:50:05.9753234Z Entering 'third_party/mimalloc' 2025-12-04T09:50:05.9775226Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T09:50:05.9795360Z Entering 'third_party/nlohmann' 2025-12-04T09:50:05.9819098Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T09:50:05.9835238Z Entering 'third_party/onnx' 2025-12-04T09:50:05.9860155Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T09:50:05.9886022Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:50:05.9909899Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:50:05.9929094Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:50:05.9950495Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T09:50:05.9967317Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:50:05.9987502Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:50:06.0002581Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:50:06.0024017Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:50:06.0041094Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:50:06.0062553Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T09:50:06.0075986Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:50:06.0097908Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T09:50:06.0117026Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:50:06.0143326Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T09:50:06.0160014Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:50:06.0179935Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T09:50:06.0195957Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:50:06.0216164Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T09:50:06.0230856Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:50:06.0252806Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T09:50:06.0271800Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:50:06.0291835Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T09:50:06.0310655Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:50:06.0330715Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T09:50:06.0361407Z Entering 'third_party/pocketfft' 2025-12-04T09:50:06.0386055Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T09:50:06.0409847Z Entering 'third_party/protobuf' 2025-12-04T09:50:06.0430930Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T09:50:06.0450847Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:50:06.0470164Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:50:06.0482611Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:50:06.0505548Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:50:06.0526295Z Entering 'third_party/psimd' 2025-12-04T09:50:06.0553319Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T09:50:06.0571301Z Entering 'third_party/pthreadpool' 2025-12-04T09:50:06.0592940Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T09:50:06.0611414Z Entering 'third_party/pybind11' 2025-12-04T09:50:06.0632038Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:50:06.0648492Z Entering 'third_party/python-peachpy' 2025-12-04T09:50:06.0671400Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T09:50:06.0691476Z Entering 'third_party/sleef' 2025-12-04T09:50:06.0712233Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T09:50:06.0727824Z Entering 'third_party/tensorpipe' 2025-12-04T09:50:06.0754442Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T09:50:06.0771386Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:50:06.0790664Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:50:06.0807972Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:50:06.0832159Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T09:50:06.0847589Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:50:06.0867280Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T09:50:06.0883929Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:50:06.0906104Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:50:06.0918647Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:50:06.0943559Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T09:50:06.0986112Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1015695Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1043636Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1077921Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1104424Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1128832Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1159596Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1191705Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1211968Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1248246Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1273045Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1303493Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1323534Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1352866Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1382783Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1412453Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1435250Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1463882Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1497701Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1528958Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1557131Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1590344Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1611870Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1639218Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1665727Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1699589Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1730473Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1753648Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1786995Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1814960Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1842164Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1878194Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1908742Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1930013Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1955911Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.1987408Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2013995Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2041775Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2074806Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2105627Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2122743Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2153166Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2193949Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2223727Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2251122Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2276805Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2303558Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2327253Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2356120Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2380948Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2407432Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2428473Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2451328Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2480066Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2510206Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2529223Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2558823Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2590472Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2615034Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2642431Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2673150Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2702363Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2727643Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2757037Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2789900Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2821451Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2856136Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2882394Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2910853Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2936863Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2974786Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.2999357Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.3021970Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.3053519Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.3078525Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.3112872Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.3143329Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.3171763Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.3199646Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.3225205Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.3258066Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:50:06.3388359Z A job completed hook has been configured by the self-hosted runner administrator 2025-12-04T09:50:06.3399332Z ##[group]Run '/home/ec2-user/runner-scripts/after_job.sh' 2025-12-04T09:50:06.3402821Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:50:06.3403091Z ##[endgroup] 2025-12-04T09:50:06.3486917Z [!ALERT!] Swap in detected! [!ALERT!] 2025-12-04T09:50:14.5410756Z [!ALERT!] Swap out detected [!ALERT!] 2025-12-04T09:50:28.3871387Z Cleaning up orphan processes