2025-09-07T07:34:57.0452717Z Current runner version: '2.328.0' 2025-09-07T07:34:57.0457273Z Runner name: 'i-05ba5cf17983f662f' 2025-09-07T07:34:57.0457870Z Runner group name: 'default' 2025-09-07T07:34:57.0458595Z Machine name: 'ip-10-0-9-178' 2025-09-07T07:34:57.0460654Z ##[group]GITHUB_TOKEN Permissions 2025-09-07T07:34:57.0462278Z Contents: read 2025-09-07T07:34:57.0462735Z Metadata: read 2025-09-07T07:34:57.0463145Z ##[endgroup] 2025-09-07T07:34:57.0464701Z Secret source: Actions 2025-09-07T07:34:57.0465284Z Prepare workflow directory 2025-09-07T07:34:57.0824468Z Prepare all required actions 2025-09-07T07:34:57.0856743Z Getting action download info 2025-09-07T07:34:57.3503275Z Download action repository 'pytorch/test-infra@main' (SHA:548a4bc624d43a01cdf165a63b041f0ae014ddbd) 2025-09-07T07:34:58.6599441Z Download action repository 'pytorch/pytorch@main' (SHA:93fb23d6fae7c4e82c4239a1033e522088742634) 2025-09-07T07:35:11.3866982Z Download action repository 'actions/setup-python@a26af69be951a213d495a4c3e4e4022e16d87065' (SHA:a26af69be951a213d495a4c3e4e4022e16d87065) 2025-09-07T07:35:11.6293633Z Download action repository 'aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722' (SHA:ececac1a45f3b08a01d2dd070d28d111c5fe6722) 2025-09-07T07:35:11.7977336Z Download action repository 'aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076' (SHA:062b18b96a7aff071d4dc91bc00c4c1a7945b076) 2025-09-07T07:35:11.9379756Z Download action repository 'seemethere/upload-artifact-s3@baba72d0712b404f646cebe0730933554ebce96a' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-09-07T07:35:12.1567576Z Getting action download info 2025-09-07T07:35:12.2557909Z Download action repository 'actions/checkout@v4' (SHA:08eba0b27e820071cde6df949e0beb9ba4906955) 2025-09-07T07:35:12.4824734Z Getting action download info 2025-09-07T07:35:12.5783805Z Download action repository 'nick-fields/retry@v3.0.0' (SHA:7152eba30c6575329ac0576536151aca5a72780e) 2025-09-07T07:35:12.8145835Z Getting action download info 2025-09-07T07:35:12.9243636Z Download action repository 'nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482' (SHA:3e91a01664abd3c5cd539100d10d33b9c5b68482) 2025-09-07T07:35:13.1781917Z Getting action download info 2025-09-07T07:35:13.3460889Z Uses: pytorch/pytorch/.github/workflows/_linux-test.yml@refs/heads/main (93fb23d6fae7c4e82c4239a1033e522088742634) 2025-09-07T07:35:13.3463676Z ##[group] Inputs 2025-09-07T07:35:13.3463974Z build-environment: linux-jammy-py3.9-gcc11-build 2025-09-07T07:35:13.3466247Z test-matrix: {"include": [{"config": "inductor_huggingface_perf_cpu_x86", "shard": 1, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_huggingface_perf_cpu_x86", "shard": 2, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_huggingface_perf_cpu_x86", "shard": 3, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 1, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 2, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 3, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 4, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 5, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 1, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 2, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 3, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 4, "num_shards": 4, "runner": "linux.24xl.spr-metal"}]} 2025-09-07T07:35:13.3469245Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:35:13.3469982Z sync-tag: 2025-09-07T07:35:13.3470612Z timeout-minutes: 720 2025-09-07T07:35:13.3470850Z use-gha: 2025-09-07T07:35:13.3471338Z dashboard-tag: training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true 2025-09-07T07:35:13.3471839Z s3-bucket: gha-artifacts 2025-09-07T07:35:13.3472058Z aws-role-to-assume: 2025-09-07T07:35:13.3472489Z disable-monitor: false 2025-09-07T07:35:13.3472758Z monitor-log-interval: 15 2025-09-07T07:35:13.3473024Z monitor-data-collect-interval: 4 2025-09-07T07:35:13.3473281Z ##[endgroup] 2025-09-07T07:35:13.3473691Z Complete job name: inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal) 2025-09-07T07:35:13.4049936Z A job started hook has been configured by the self-hosted runner administrator 2025-09-07T07:35:13.4233684Z ##[group]Run '/home/ec2-user/runner-scripts/before_job.sh' 2025-09-07T07:35:13.4239567Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:35:13.4239963Z ##[endgroup] 2025-09-07T07:35:14.3414851Z Runner Type: linux.24xl.spr-metal 2025-09-07T07:35:14.3415248Z Instance Type: c7i.metal-24xl 2025-09-07T07:35:14.3415450Z AMI Name: unknown 2025-09-07T07:35:14.3447061Z AMI ID: ami-05ffe3c48a9991133 2025-09-07T07:35:18.0099106Z ##[group]Run pytorch/test-infra/.github/actions/setup-ssh@main 2025-09-07T07:35:18.0099427Z with: 2025-09-07T07:35:18.0099931Z github-secret: *** 2025-09-07T07:35:18.0100367Z instructions: All testing is done inside the container, to start an interactive session run: docker exec -it $(docker container ps --format '{{.ID}}') bash 2025-09-07T07:35:18.0100808Z activate-with-label: false 2025-09-07T07:35:18.0100987Z label: with-ssh 2025-09-07T07:35:18.0101158Z remove-existing-keys: true 2025-09-07T07:35:18.0101339Z fail-silently: true 2025-09-07T07:35:18.0101558Z env: 2025-09-07T07:35:18.0101697Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:35:18.0101888Z ##[endgroup] 2025-09-07T07:35:18.1135826Z Please see https://github.com/pytorch/pytorch/wiki/Debugging-using-with-ssh-for-Github-Actions for more info. 2025-09-07T07:35:18.1136408Z Not on pull request and ciflow reference could not be extracted, skipping adding ssh keys 2025-09-07T07:35:18.1252086Z ##[group]Run pytorch/pytorch/.github/actions/checkout-pytorch@main 2025-09-07T07:35:18.1252356Z with: 2025-09-07T07:35:18.1252521Z no-sudo: true 2025-09-07T07:35:18.1252689Z submodules: recursive 2025-09-07T07:35:18.1252860Z fetch-depth: 0 2025-09-07T07:35:18.1253012Z env: 2025-09-07T07:35:18.1253162Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:35:18.1253337Z ##[endgroup] 2025-09-07T07:35:18.1307055Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:35:18.1307602Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:35:18.1315123Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:35:18.1315381Z env: 2025-09-07T07:35:18.1315562Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:35:18.1315772Z ##[endgroup] 2025-09-07T07:35:18.1375230Z ##[group]Run # Use all available CPUs for fetching 2025-09-07T07:35:18.1375538Z # Use all available CPUs for fetching 2025-09-07T07:35:18.1375767Z cd "${GITHUB_WORKSPACE}" 2025-09-07T07:35:18.1375989Z git config --global fetch.parallel 0 2025-09-07T07:35:18.1376241Z git config --global submodule.fetchJobs 0 2025-09-07T07:35:18.1376462Z  2025-09-07T07:35:18.1376694Z # Clean workspace. The default checkout action should also do this, but 2025-09-07T07:35:18.1376976Z # do it here as well just in case 2025-09-07T07:35:18.1377188Z if [[ -d .git ]]; then 2025-09-07T07:35:18.1377389Z  if [ -z "${NO_SUDO}" ]; then 2025-09-07T07:35:18.1377591Z  sudo git clean -ffdx 2025-09-07T07:35:18.1377769Z  else 2025-09-07T07:35:18.1378053Z  git clean -ffdx 2025-09-07T07:35:18.1378226Z  fi 2025-09-07T07:35:18.1378372Z fi 2025-09-07T07:35:18.1382366Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:35:18.1382610Z env: 2025-09-07T07:35:18.1382854Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:35:18.1383054Z NO_SUDO: true 2025-09-07T07:35:18.1383212Z ##[endgroup] 2025-09-07T07:35:18.1478249Z ##[group]Run actions/checkout@v4 2025-09-07T07:35:18.1478470Z with: 2025-09-07T07:35:18.1478641Z ref: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:35:18.1478847Z fetch-depth: 0 2025-09-07T07:35:18.1479015Z submodules: recursive 2025-09-07T07:35:18.1479187Z show-progress: false 2025-09-07T07:35:18.1479361Z repository: pytorch/pytorch 2025-09-07T07:35:18.1479613Z token: *** 2025-09-07T07:35:18.1479767Z ssh-strict: true 2025-09-07T07:35:18.1479924Z ssh-user: git 2025-09-07T07:35:18.1480082Z persist-credentials: true 2025-09-07T07:35:18.1480251Z clean: true 2025-09-07T07:35:18.1480433Z sparse-checkout-cone-mode: true 2025-09-07T07:35:18.1480624Z fetch-tags: false 2025-09-07T07:35:18.1480782Z lfs: false 2025-09-07T07:35:18.1480924Z set-safe-directory: true 2025-09-07T07:35:18.1481105Z env: 2025-09-07T07:35:18.1481250Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:35:18.1481420Z ##[endgroup] 2025-09-07T07:35:18.2268797Z Syncing repository: pytorch/pytorch 2025-09-07T07:35:18.2269788Z ##[group]Getting Git version info 2025-09-07T07:35:18.2270100Z Working directory is '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-09-07T07:35:18.2270620Z [command]/usr/bin/git version 2025-09-07T07:35:18.2464451Z git version 2.47.1 2025-09-07T07:35:18.2479778Z ##[endgroup] 2025-09-07T07:35:18.2488339Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/a851cf05-a7f7-4fe3-85c6-4c81df138e2f/.gitconfig' 2025-09-07T07:35:18.2501556Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/a851cf05-a7f7-4fe3-85c6-4c81df138e2f' before making global git config changes 2025-09-07T07:35:18.2502188Z Adding repository directory to the temporary git global config as a safe directory 2025-09-07T07:35:18.2505326Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T07:35:18.2540000Z Deleting the contents of '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-09-07T07:35:18.2542203Z ##[group]Initializing the repository 2025-09-07T07:35:18.2545179Z [command]/usr/bin/git init /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T07:35:18.2590681Z hint: Using 'master' as the name for the initial branch. This default branch name 2025-09-07T07:35:18.2591130Z hint: is subject to change. To configure the initial branch name to use in all 2025-09-07T07:35:18.2591564Z hint: of your new repositories, which will suppress this warning, call: 2025-09-07T07:35:18.2591862Z hint: 2025-09-07T07:35:18.2592109Z hint: git config --global init.defaultBranch 2025-09-07T07:35:18.2592403Z hint: 2025-09-07T07:35:18.2592678Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2025-09-07T07:35:18.2593077Z hint: 'development'. The just-created branch can be renamed via this command: 2025-09-07T07:35:18.2593397Z hint: 2025-09-07T07:35:18.2593591Z hint: git branch -m 2025-09-07T07:35:18.2610367Z Initialized empty Git repository in /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/ 2025-09-07T07:35:18.2616595Z [command]/usr/bin/git remote add origin https://github.com/pytorch/pytorch 2025-09-07T07:35:18.2645605Z ##[endgroup] 2025-09-07T07:35:18.2645945Z ##[group]Disabling automatic garbage collection 2025-09-07T07:35:18.2648513Z [command]/usr/bin/git config --local gc.auto 0 2025-09-07T07:35:18.2669075Z ##[endgroup] 2025-09-07T07:35:18.2669359Z ##[group]Setting up auth 2025-09-07T07:35:18.2673914Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-09-07T07:35:18.2692136Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-09-07T07:35:18.2967466Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-09-07T07:35:18.2989871Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-09-07T07:35:18.3240458Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-09-07T07:35:18.3282072Z ##[endgroup] 2025-09-07T07:35:18.3282400Z ##[group]Fetching the repository 2025-09-07T07:35:18.3287546Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2025-09-07T07:35:59.9285243Z From https://github.com/pytorch/pytorch 2025-09-07T07:35:59.9285628Z * [new branch] 160583 -> origin/160583 2025-09-07T07:35:59.9285979Z * [new branch] 2.6.0.dev20241004+ -> origin/2.6.0.dev20241004+ 2025-09-07T07:35:59.9286296Z * [new branch] 5addvllmbuild -> origin/5addvllmbuild 2025-09-07T07:35:59.9286645Z * [new branch] AaronWang04_addmmfusion_perftest -> origin/AaronWang04_addmmfusion_perftest 2025-09-07T07:35:59.9287042Z * [new branch] HDCharles-2.6.0-release-notes -> origin/HDCharles-2.6.0-release-notes 2025-09-07T07:35:59.9287374Z * [new branch] ISSUE-154849 -> origin/ISSUE-154849 2025-09-07T07:35:59.9287747Z * [new branch] JackCaoG/dynamo_make_fx_non_core_aten_ops -> origin/JackCaoG/dynamo_make_fx_non_core_aten_ops 2025-09-07T07:35:59.9288122Z * [new branch] NicoshevSVE128 -> origin/NicoshevSVE128 2025-09-07T07:35:59.9288613Z * [new branch] PR-AOTInductorNoneBug -> origin/PR-AOTInductorNoneBug 2025-09-07T07:35:59.9288977Z * [new branch] PR-AOTInductorNoneBugFix -> origin/PR-AOTInductorNoneBugFix 2025-09-07T07:35:59.9289641Z * [new branch] PR-FixConfigsIssue -> origin/PR-FixConfigsIssue 2025-09-07T07:35:59.9290016Z * [new branch] PR-NoneBugFix-viable -> origin/PR-NoneBugFix-viable 2025-09-07T07:35:59.9290340Z * [new branch] PR-ResetToZero -> origin/PR-ResetToZero 2025-09-07T07:35:59.9290671Z * [new branch] Update-Flash-Packaging -> origin/Update-Flash-Packaging 2025-09-07T07:35:59.9290983Z * [new branch] VLA_exp -> origin/VLA_exp 2025-09-07T07:35:59.9291309Z * [new branch] actually-run-mps-aot-inductor -> origin/actually-run-mps-aot-inductor 2025-09-07T07:35:59.9291714Z * [new branch] add-missing-args-normalization -> origin/add-missing-args-normalization 2025-09-07T07:35:59.9292089Z * [new branch] add-user-guide-structure -> origin/add-user-guide-structure 2025-09-07T07:35:59.9292436Z * [new branch] add-vllm-nightly-build -> origin/add-vllm-nightly-build 2025-09-07T07:35:59.9292778Z * [new branch] add_compile_benchmarking -> origin/add_compile_benchmarking 2025-09-07T07:35:59.9293103Z * [new branch] addmm-heuristic -> origin/addmm-heuristic 2025-09-07T07:35:59.9293538Z * [new branch] addsimde -> origin/addsimde 2025-09-07T07:35:59.9294070Z * [new branch] addvllmtest -> origin/addvllmtest 2025-09-07T07:35:59.9295071Z * [new branch] adi/acl_upgrade -> origin/adi/acl_upgrade 2025-09-07T07:35:59.9295617Z * [new branch] adi/test -> origin/adi/test 2025-09-07T07:35:59.9296411Z * [new branch] adi/test_bgemm -> origin/adi/test_bgemm 2025-09-07T07:35:59.9296712Z * [new branch] adi/test_fusions -> origin/adi/test_fusions 2025-09-07T07:35:59.9297403Z * [new branch] adi/test_onednn_v3.9 -> origin/adi/test_onednn_v3.9 2025-09-07T07:35:59.9298182Z * [new branch] adi/test_presve_change -> origin/adi/test_presve_change 2025-09-07T07:35:59.9298499Z * [new branch] adi/test_timm -> origin/adi/test_timm 2025-09-07T07:35:59.9299520Z * [new branch] adi/testpresve_change -> origin/adi/testpresve_change 2025-09-07T07:35:59.9300743Z * [new branch] aditew01/test/vec_bf16 -> origin/aditew01/test/vec_bf16 2025-09-07T07:35:59.9301358Z * [new branch] ah-globalfeedback-hook -> origin/ah-globalfeedback-hook 2025-09-07T07:35:59.9301923Z * [new branch] alt-disable -> origin/alt-disable 2025-09-07T07:35:59.9302941Z * [new branch] angelayi/aoti_additional_files -> origin/angelayi/aoti_additional_files 2025-09-07T07:35:59.9303491Z * [new branch] angelayi/aoti_inductor_fx -> origin/angelayi/aoti_inductor_fx 2025-09-07T07:35:59.9304051Z * [new branch] angelayi/benchmark -> origin/angelayi/benchmark 2025-09-07T07:35:59.9304764Z * [new branch] angelayi/benchmark2 -> origin/angelayi/benchmark2 2025-09-07T07:35:59.9305287Z * [new branch] angelayi/change_pytree_serialization -> origin/angelayi/change_pytree_serialization 2025-09-07T07:35:59.9305846Z * [new branch] angelayi/cpp_loader -> origin/angelayi/cpp_loader 2025-09-07T07:35:59.9306706Z * [new branch] angelayi/custom_op_subgraph -> origin/angelayi/custom_op_subgraph 2025-09-07T07:35:59.9307284Z * [new branch] angelayi/customop -> origin/angelayi/customop 2025-09-07T07:35:59.9308180Z * [new branch] angelayi/fake_cache_empty -> origin/angelayi/fake_cache_empty 2025-09-07T07:35:59.9308714Z * [new branch] angelayi/is_symbolic_tracing -> origin/angelayi/is_symbolic_tracing 2025-09-07T07:35:59.9309289Z * [new branch] angelayi/item -> origin/angelayi/item 2025-09-07T07:35:59.9310324Z * [new branch] angelayi/no_so_weight -> origin/angelayi/no_so_weight 2025-09-07T07:35:59.9310654Z * [new branch] angelayi/opoverload -> origin/angelayi/opoverload 2025-09-07T07:35:59.9311143Z * [new branch] angelayi/pattern -> origin/angelayi/pattern 2025-09-07T07:35:59.9311725Z * [new branch] angelayi/pytree -> origin/angelayi/pytree 2025-09-07T07:35:59.9312332Z * [new branch] angelayi/scan_layers -> origin/angelayi/scan_layers 2025-09-07T07:35:59.9312929Z * [new branch] angelayi/symint_input -> origin/angelayi/symint_input 2025-09-07T07:35:59.9313438Z * [new branch] angelayi/test_cpp -> origin/angelayi/test_cpp 2025-09-07T07:35:59.9313984Z * [new branch] angelayi/torch_size -> origin/angelayi/torch_size 2025-09-07T07:35:59.9314587Z * [new branch] aoti-cuda-alloc -> origin/aoti-cuda-alloc 2025-09-07T07:35:59.9315191Z * [new branch] aoti_target_windows -> origin/aoti_target_windows 2025-09-07T07:35:59.9315765Z * [new branch] aoti_weight_sharing -> origin/aoti_weight_sharing 2025-09-07T07:35:59.9316526Z * [new branch] atalman-inductor-perf-cu124 -> origin/atalman-inductor-perf-cu124 2025-09-07T07:35:59.9317042Z * [new branch] atalman-inductor-perf-cu124.1 -> origin/atalman-inductor-perf-cu124.1 2025-09-07T07:35:59.9317668Z * [new branch] atalman-patch-1 -> origin/atalman-patch-1 2025-09-07T07:35:59.9318412Z * [new branch] atalman-patch-3 -> origin/atalman-patch-3 2025-09-07T07:35:59.9319128Z * [new branch] atalman-patch-4 -> origin/atalman-patch-4 2025-09-07T07:35:59.9319709Z * [new branch] atalman-patch-5 -> origin/atalman-patch-5 2025-09-07T07:35:59.9320306Z * [new branch] atalman-patch-6 -> origin/atalman-patch-6 2025-09-07T07:35:59.9321041Z * [new branch] atalman_inductor_2.3.0 -> origin/atalman_inductor_2.3.0 2025-09-07T07:35:59.9321544Z * [new branch] atalman_inductor_2.3.1 -> origin/atalman_inductor_2.3.1 2025-09-07T07:35:59.9322092Z * [new branch] atalman_inductor_2.4.0 -> origin/atalman_inductor_2.4.0 2025-09-07T07:35:59.9322683Z * [new branch] atalman_inductor_2.4.x -> origin/atalman_inductor_2.4.x 2025-09-07T07:35:59.9323332Z * [new branch] autoupdate-transformers-pin-via-pr -> origin/autoupdate-transformers-pin-via-pr 2025-09-07T07:35:59.9324149Z * [new branch] bahuang/dtensor_demo -> origin/bahuang/dtensor_demo 2025-09-07T07:35:59.9324641Z * [new branch] bahuang/test -> origin/bahuang/test 2025-09-07T07:35:59.9325724Z * [new branch] base/1.5 -> origin/base/1.5 2025-09-07T07:35:59.9326311Z * [new branch] batching_sdpa_efficient_attention -> origin/batching_sdpa_efficient_attention 2025-09-07T07:35:59.9326861Z * [new branch] bc-lint-config -> origin/bc-lint-config 2025-09-07T07:35:59.9327479Z * [new branch] bc-lint-test-new-config -> origin/bc-lint-test-new-config 2025-09-07T07:35:59.9328326Z * [new branch] benchmark-updates -> origin/benchmark-updates 2025-09-07T07:35:59.9329376Z * [new branch] benchmarker_compat_with_do_bench -> origin/benchmarker_compat_with_do_bench 2025-09-07T07:35:59.9329883Z * [new branch] benchmarking-script -> origin/benchmarking-script 2025-09-07T07:35:59.9330804Z * [new branch] bertmaher/pinbump26 -> origin/bertmaher/pinbump26 2025-09-07T07:35:59.9331660Z * [new branch] bertrand/cutlass -> origin/bertrand/cutlass 2025-09-07T07:35:59.9332508Z * [new branch] bf/cg-custom-wrapper -> origin/bf/cg-custom-wrapper 2025-09-07T07:35:59.9333036Z * [new branch] bf/cg-or-error -> origin/bf/cg-or-error 2025-09-07T07:35:59.9333507Z * [new branch] bf/cg-remove-check -> origin/bf/cg-remove-check 2025-09-07T07:35:59.9334040Z * [new branch] bf/cg-skip-1-kernel -> origin/bf/cg-skip-1-kernel 2025-09-07T07:35:59.9334605Z * [new branch] bf/cudagraph -> origin/bf/cudagraph 2025-09-07T07:35:59.9335458Z * [new branch] bf/cudagraph-disable-input-mutation -> origin/bf/cudagraph-disable-input-mutation 2025-09-07T07:35:59.9336460Z * [new branch] bf/cudagraph-enable-input-mutation-support-benchmark -> origin/bf/cudagraph-enable-input-mutation-support-benchmark 2025-09-07T07:35:59.9336961Z * [new branch] bf/cudagraph-partition -> origin/bf/cudagraph-partition 2025-09-07T07:35:59.9337433Z * [new branch] bf/default-recompile-reason -> origin/bf/default-recompile-reason 2025-09-07T07:35:59.9337975Z * [new branch] bf/donated-buffer-bench -> origin/bf/donated-buffer-bench 2025-09-07T07:35:59.9338639Z * [new branch] bf/exp -> origin/bf/exp 2025-09-07T07:35:59.9339164Z * [new branch] bf/pa-non-divisible -> origin/bf/pa-non-divisible 2025-09-07T07:35:59.9339720Z * [new branch] bf/partition-move-cpu -> origin/bf/partition-move-cpu 2025-09-07T07:35:59.9340292Z * [new branch] bf/partition-turn-on -> origin/bf/partition-turn-on 2025-09-07T07:35:59.9340870Z * [new branch] bf/remove-check-55b0c39d -> origin/bf/remove-check-55b0c39d 2025-09-07T07:35:59.9341390Z * [new branch] bf/rope -> origin/bf/rope 2025-09-07T07:35:59.9342041Z * [new branch] bisect_perf_hf_T5_3acc6eac492 -> origin/bisect_perf_hf_T5_3acc6eac492 2025-09-07T07:35:59.9342552Z * [new branch] bisect_perf_hf_T5_3fcf66f61fb -> origin/bisect_perf_hf_T5_3fcf66f61fb 2025-09-07T07:35:59.9343068Z * [new branch] bisect_perf_hf_T5_4009d154129 -> origin/bisect_perf_hf_T5_4009d154129 2025-09-07T07:35:59.9343570Z * [new branch] bisect_perf_hf_T5_40d0740e73d -> origin/bisect_perf_hf_T5_40d0740e73d 2025-09-07T07:35:59.9344080Z * [new branch] bisect_perf_hf_T5_5268754e -> origin/bisect_perf_hf_T5_5268754e 2025-09-07T07:35:59.9344596Z * [new branch] bisect_perf_hf_T5_7d89a8d385c -> origin/bisect_perf_hf_T5_7d89a8d385c 2025-09-07T07:35:59.9345120Z * [new branch] bisect_perf_hf_T5_b7a25c1ee7c -> origin/bisect_perf_hf_T5_b7a25c1ee7c 2025-09-07T07:35:59.9345643Z * [new branch] bisect_perf_hf_T5_c25b201583f -> origin/bisect_perf_hf_T5_c25b201583f 2025-09-07T07:35:59.9346175Z * [new branch] bisect_perf_hf_T5_c93e57efac0 -> origin/bisect_perf_hf_T5_c93e57efac0 2025-09-07T07:35:59.9346947Z * [new branch] bisect_perf_hf_T5_ca9813ea149 -> origin/bisect_perf_hf_T5_ca9813ea149 2025-09-07T07:35:59.9347404Z * [new branch] bisect_perf_hf_T5_d65f194a -> origin/bisect_perf_hf_T5_d65f194a 2025-09-07T07:35:59.9347933Z * [new branch] bisect_perf_hf_T5_da94ab0b -> origin/bisect_perf_hf_T5_da94ab0b 2025-09-07T07:35:59.9348492Z * [new branch] bisect_perf_hf_T5_da94ab0b_new -> origin/bisect_perf_hf_T5_da94ab0b_new 2025-09-07T07:35:59.9348983Z * [new branch] bisect_perf_hf_T5_db4e8a1d8a8 -> origin/bisect_perf_hf_T5_db4e8a1d8a8 2025-09-07T07:35:59.9349487Z * [new branch] bisect_perf_hf_T5_e0d97e936a2 -> origin/bisect_perf_hf_T5_e0d97e936a2 2025-09-07T07:35:59.9350032Z * [new branch] bisect_perf_hf_T5_f23621ec563 -> origin/bisect_perf_hf_T5_f23621ec563 2025-09-07T07:35:59.9351044Z * [new branch] bowbao/bench_updates_stage -> origin/bowbao/bench_updates_stage 2025-09-07T07:35:59.9351531Z * [new branch] bowbao/dort_rewriter -> origin/bowbao/dort_rewriter 2025-09-07T07:35:59.9352099Z * [new branch] bowbao/wip_prs -> origin/bowbao/wip_prs 2025-09-07T07:35:59.9352983Z * [new branch] brister/break_tensorbox -> origin/brister/break_tensorbox 2025-09-07T07:35:59.9353434Z * [new branch] brister/custom_fx_backend -> origin/brister/custom_fx_backend 2025-09-07T07:35:59.9353967Z * [new branch] brister/fx_custom_triton -> origin/brister/fx_custom_triton 2025-09-07T07:35:59.9354483Z * [new branch] brister/tensor_box_output -> origin/brister/tensor_box_output 2025-09-07T07:35:59.9355050Z * [new branch] brister/tiled_reduction_no_numel_check -> origin/brister/tiled_reduction_no_numel_check 2025-09-07T07:35:59.9355902Z * [new branch] c57382a49 -> origin/c57382a49 2025-09-07T07:35:59.9356386Z * [new branch] ca_0431d47eaa -> origin/ca_0431d47eaa 2025-09-07T07:35:59.9356885Z * [new branch] ca_fix_0431d47eaa -> origin/ca_fix_0431d47eaa 2025-09-07T07:35:59.9358196Z * [new branch] camyll/revert-94bc900da97ad7f3c35b3b819bb53b23c74b581a-for-release-2.8 -> origin/camyll/revert-94bc900da97ad7f3c35b3b819bb53b23c74b581a-for-release-2.8 2025-09-07T07:35:59.9358801Z * [new branch] camyllh/test_setup_hooks_push -> origin/camyllh/test_setup_hooks_push 2025-09-07T07:35:59.9359534Z * [new branch] cherry-pick-149654-by-pytorch_bot_bot_ -> origin/cherry-pick-149654-by-pytorch_bot_bot_ 2025-09-07T07:35:59.9360089Z * [new branch] cherry-pick-151939-by-pytorch_bot_bot_ -> origin/cherry-pick-151939-by-pytorch_bot_bot_ 2025-09-07T07:35:59.9360681Z * [new branch] cherry-pick-154174-by-pytorch_bot_bot_ -> origin/cherry-pick-154174-by-pytorch_bot_bot_ 2025-09-07T07:35:59.9361295Z * [new branch] cherry-pick-156260-by-pytorch_bot_bot_ -> origin/cherry-pick-156260-by-pytorch_bot_bot_ 2025-09-07T07:35:59.9361867Z * [new branch] cherry-pick-157453-by-pytorch_bot_bot_ -> origin/cherry-pick-157453-by-pytorch_bot_bot_ 2025-09-07T07:35:59.9362500Z * [new branch] cherry-pick-157513-by-pytorch_bot_bot_ -> origin/cherry-pick-157513-by-pytorch_bot_bot_ 2025-09-07T07:35:59.9363103Z * [new branch] cherry-pick-157695-by-pytorch_bot_bot_ -> origin/cherry-pick-157695-by-pytorch_bot_bot_ 2025-09-07T07:35:59.9363679Z * [new branch] cherry-pick-157732-by-pytorch_bot_bot_ -> origin/cherry-pick-157732-by-pytorch_bot_bot_ 2025-09-07T07:35:59.9364238Z * [new branch] cherry-pick-158537-by-pytorch_bot_bot_ -> origin/cherry-pick-158537-by-pytorch_bot_bot_ 2025-09-07T07:35:59.9364820Z * [new branch] cherry-pick-159969-by-pytorch_bot_bot_ -> origin/cherry-pick-159969-by-pytorch_bot_bot_ 2025-09-07T07:35:59.9365551Z * [new branch] cherry-pick-160586-by-pytorch_bot_bot_ -> origin/cherry-pick-160586-by-pytorch_bot_bot_ 2025-09-07T07:35:59.9366504Z * [new branch] chilli/flex_vllm -> origin/chilli/flex_vllm 2025-09-07T07:35:59.9367119Z * [new branch] cleanup-inductor-benchmark-images -> origin/cleanup-inductor-benchmark-images 2025-09-07T07:35:59.9367623Z * [new branch] codex-testing -> origin/codex-testing 2025-09-07T07:35:59.9368881Z * [new branch] codex/add-helper-function-to-sizevars.py -> origin/codex/add-helper-function-to-sizevars.py 2025-09-07T07:35:59.9369417Z * [new branch] codex/add-helper-function-to-sizevars.py_2025-09-05 -> origin/codex/add-helper-function-to-sizevars.py_2025-09-05 2025-09-07T07:35:59.9369938Z * [new branch] codex/add-metadata-field-for-file-path -> origin/codex/add-metadata-field-for-file-path 2025-09-07T07:35:59.9370723Z * [new branch] codex/add-test-for-inductor-local-cache-behavior -> origin/codex/add-test-for-inductor-local-cache-behavior 2025-09-07T07:35:59.9371572Z * [new branch] codex/create-test-for-tensor-memory-leak-in-cudagraph -> origin/codex/create-test-for-tensor-memory-leak-in-cudagraph 2025-09-07T07:35:59.9372097Z * [new branch] codex/fix-issue-121219-in-pytorch -> origin/codex/fix-issue-121219-in-pytorch 2025-09-07T07:35:59.9372559Z * [new branch] codex/fix-issue-160415-in-pytorch -> origin/codex/fix-issue-160415-in-pytorch 2025-09-07T07:35:59.9373204Z * [new branch] codex/fix-noqengine-quantized-engine-support -> origin/codex/fix-noqengine-quantized-engine-support 2025-09-07T07:35:59.9373710Z * [new branch] codex/fix-pin_memory-error-handling -> origin/codex/fix-pin_memory-error-handling 2025-09-07T07:35:59.9374249Z * [new branch] codex/propose-fix-for-issue-160332 -> origin/codex/propose-fix-for-issue-160332 2025-09-07T07:35:59.9375045Z * [new branch] codex/refactor-lintrunner-config-to-use-uv-run -> origin/codex/refactor-lintrunner-config-to-use-uv-run 2025-09-07T07:35:59.9376221Z * [new branch] codex/remove-allow-untyped-defs-and-fix-type-errors -> origin/codex/remove-allow-untyped-defs-and-fix-type-errors 2025-09-07T07:35:59.9376820Z * [new branch] compile_fsdp2_disable_stream_and_event -> origin/compile_fsdp2_disable_stream_and_event 2025-09-07T07:35:59.9377207Z * [new branch] context_test -> origin/context_test 2025-09-07T07:35:59.9377594Z * [new branch] copilot/fix-157446 -> origin/copilot/fix-157446 2025-09-07T07:35:59.9378039Z * [new branch] copy_graph -> origin/copy_graph 2025-09-07T07:35:59.9378965Z * [new branch] cpio/fix_new_ami_tests -> origin/cpio/fix_new_ami_tests 2025-09-07T07:35:59.9379839Z * [new branch] csl/always_produce_xml -> origin/csl/always_produce_xml 2025-09-07T07:35:59.9380310Z * [new branch] csl/build_test_more_procs -> origin/csl/build_test_more_procs 2025-09-07T07:35:59.9380860Z * [new branch] csl/build_test_more_procs2 -> origin/csl/build_test_more_procs2 2025-09-07T07:35:59.9381410Z * [new branch] csl/disable_flaky_cpp_test -> origin/csl/disable_flaky_cpp_test 2025-09-07T07:35:59.9381871Z * [new branch] csl/disable_periodic_test -> origin/csl/disable_periodic_test 2025-09-07T07:35:59.9382649Z * [new branch] csl/exclude_rocm_viable_strict -> origin/csl/exclude_rocm_viable_strict 2025-09-07T07:35:59.9383292Z * [new branch] csl/katex -> origin/csl/katex 2025-09-07T07:35:59.9383832Z * [new branch] csl/larger_runner -> origin/csl/larger_runner 2025-09-07T07:35:59.9384563Z * [new branch] csl/lintrunner_stuff -> origin/csl/lintrunner_stuff 2025-09-07T07:35:59.9385052Z * [new branch] csl/mps_sharding -> origin/csl/mps_sharding 2025-09-07T07:35:59.9385584Z * [new branch] csl/multistage_docker -> origin/csl/multistage_docker 2025-09-07T07:35:59.9386204Z * [new branch] csl/name_link_check_job -> origin/csl/name_link_check_job 2025-09-07T07:35:59.9386742Z * [new branch] csl/no_keep_goin_rocm -> origin/csl/no_keep_goin_rocm 2025-09-07T07:35:59.9387332Z * [new branch] csl/not_600_timeout -> origin/csl/not_600_timeout 2025-09-07T07:35:59.9387842Z * [new branch] csl/revert_open -> origin/csl/revert_open 2025-09-07T07:35:59.9388374Z * [new branch] csl/skip_build -> origin/csl/skip_build 2025-09-07T07:35:59.9388941Z * [new branch] csl/test_cuda_build_large_runner -> origin/csl/test_cuda_build_large_runner 2025-09-07T07:35:59.9389588Z * [new branch] csl/win_sccache -> origin/csl/win_sccache 2025-09-07T07:35:59.9390150Z * [new branch] cublasltrelax2 -> origin/cublasltrelax2 2025-09-07T07:35:59.9390725Z * [new branch] cublasrelax2 -> origin/cublasrelax2 2025-09-07T07:35:59.9391339Z * [new branch] cudnnsdparefactor -> origin/cudnnsdparefactor 2025-09-07T07:35:59.9391867Z * [new branch] custom_lowering_dict -> origin/custom_lowering_dict 2025-09-07T07:35:59.9392359Z * [new branch] czhuge_muon_dev -> origin/czhuge_muon_dev 2025-09-07T07:35:59.9393458Z * [new branch] d4l3k/delete_hook -> origin/d4l3k/delete_hook 2025-09-07T07:35:59.9393912Z * [new branch] dcp_zoc -> origin/dcp_zoc 2025-09-07T07:35:59.9394471Z * [new branch] debug-guard -> origin/debug-guard 2025-09-07T07:35:59.9395057Z * [new branch] delete-quant-docs -> origin/delete-quant-docs 2025-09-07T07:35:59.9397511Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.2 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.2 2025-09-07T07:35:59.9398365Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.3 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.3 2025-09-07T07:35:59.9399090Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.4 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.4 2025-09-07T07:35:59.9399769Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.56.0 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.56.0 2025-09-07T07:35:59.9400378Z * [new branch] dependabot/pip/dot-ci/docker/protobuf-5.29.5 -> origin/dependabot/pip/dot-ci/docker/protobuf-5.29.5 2025-09-07T07:35:59.9401416Z * [new branch] dependabot/pip/dot-github/requirements/protobuf-5.29.5 -> origin/dependabot/pip/dot-github/requirements/protobuf-5.29.5 2025-09-07T07:35:59.9402101Z * [new branch] desertfire/test_cpp_wrapper -> origin/desertfire/test_cpp_wrapper 2025-09-07T07:35:59.9402674Z * [new branch] desertfire/triton-cpu-for-aarch64 -> origin/desertfire/triton-cpu-for-aarch64 2025-09-07T07:35:59.9404113Z * [new branch] dev/joona/MPSNDArrayAdd -> origin/dev/joona/MPSNDArrayAdd 2025-09-07T07:35:59.9404740Z * [new branch] dev/joona/Unranked -> origin/dev/joona/Unranked 2025-09-07T07:35:59.9405623Z * [new branch] dev/joona/cat -> origin/dev/joona/cat 2025-09-07T07:35:59.9406449Z * [new branch] dev/joona/cat_remove_graph -> origin/dev/joona/cat_remove_graph 2025-09-07T07:35:59.9406947Z * [new branch] dev/joona/embeddingbag -> origin/dev/joona/embeddingbag 2025-09-07T07:35:59.9407798Z * [new branch] dev/joona/getTensorsString -> origin/dev/joona/getTensorsString 2025-09-07T07:35:59.9408753Z * [new branch] dev/joona/maxpool2dwithindices_errmsg -> origin/dev/joona/maxpool2dwithindices_errmsg 2025-09-07T07:35:59.9409668Z * [new branch] dev/joona/mps_linear_macos14 -> origin/dev/joona/mps_linear_macos14 2025-09-07T07:35:59.9410523Z * [new branch] dev/joona/sdpa -> origin/dev/joona/sdpa 2025-09-07T07:35:59.9411341Z * [new branch] dev/joona/topk_newapi -> origin/dev/joona/topk_newapi 2025-09-07T07:35:59.9411942Z * [new branch] dev/joona/type_inf -> origin/dev/joona/type_inf 2025-09-07T07:35:59.9412731Z * [new branch] dev/joona/upsize3d -> origin/dev/joona/upsize3d 2025-09-07T07:35:59.9413177Z * [new branch] disable -> origin/disable 2025-09-07T07:35:59.9413717Z * [new branch] e2e-baseline -> origin/e2e-baseline 2025-09-07T07:35:59.9414318Z * [new branch] eigen_for_sparse_addmm_v2 -> origin/eigen_for_sparse_addmm_v2 2025-09-07T07:35:59.9415218Z * [new branch] embg/test_inductor_ci_128B -> origin/embg/test_inductor_ci_128B 2025-09-07T07:35:59.9415768Z * [new branch] embg/test_inductor_ci_base -> origin/embg/test_inductor_ci_base 2025-09-07T07:35:59.9416297Z * [new branch] embg/test_inductor_ci_control -> origin/embg/test_inductor_ci_control 2025-09-07T07:35:59.9416944Z * [new branch] embg/triton_l2_prefetch_128B -> origin/embg/triton_l2_prefetch_128B 2025-09-07T07:35:59.9417565Z * [new branch] embg/triton_l2_prefetch_256B -> origin/embg/triton_l2_prefetch_256B 2025-09-07T07:35:59.9418294Z * [new branch] eqy-patch-1 -> origin/eqy-patch-1 2025-09-07T07:35:59.9418763Z * [new branch] eqy-patch-2 -> origin/eqy-patch-2 2025-09-07T07:35:59.9419270Z * [new branch] eqy-patch-3 -> origin/eqy-patch-3 2025-09-07T07:35:59.9419791Z * [new branch] eqy-patch-4 -> origin/eqy-patch-4 2025-09-07T07:35:59.9420394Z * [new branch] example-convert-torch.nn -> origin/example-convert-torch.nn 2025-09-07T07:35:59.9421408Z * [new branch] exclamaforte/add-contiguous-threshold -> origin/exclamaforte/add-contiguous-threshold 2025-09-07T07:35:59.9421820Z * [new branch] exclamaforte/amd-ma -> origin/exclamaforte/amd-ma 2025-09-07T07:35:59.9422407Z * [new branch] exclamaforte/bump-transformer-version -> origin/exclamaforte/bump-transformer-version 2025-09-07T07:35:59.9422950Z * [new branch] exclamaforte/clear-feedback-savers -> origin/exclamaforte/clear-feedback-savers 2025-09-07T07:35:59.9423505Z * [new branch] exclamaforte/combo-kernels-perf-run -> origin/exclamaforte/combo-kernels-perf-run 2025-09-07T07:35:59.9424192Z * [new branch] exclamaforte/do_bench_refactor -> origin/exclamaforte/do_bench_refactor 2025-09-07T07:35:59.9425084Z * [new branch] exclamaforte/enable-mem-dep-fusion -> origin/exclamaforte/enable-mem-dep-fusion 2025-09-07T07:35:59.9425731Z * [new branch] exclamaforte/fix-exhaustive-autotuning -> origin/exclamaforte/fix-exhaustive-autotuning 2025-09-07T07:35:59.9426392Z * [new branch] exclamaforte/fix-exhuastive-autotuning-reland -> origin/exclamaforte/fix-exhuastive-autotuning-reland 2025-09-07T07:35:59.9426894Z * [new branch] exclamaforte/fix-trace-parsing-fx-svg -> origin/exclamaforte/fix-trace-parsing-fx-svg 2025-09-07T07:35:59.9427462Z * [new branch] exclamaforte/force-pointwise-cat-perf-run -> origin/exclamaforte/force-pointwise-cat-perf-run 2025-09-07T07:35:59.9427949Z * [new branch] exclamaforte/fusion-data -> origin/exclamaforte/fusion-data 2025-09-07T07:35:59.9428553Z * [new branch] exclamaforte/gemm-benchmark-run -> origin/exclamaforte/gemm-benchmark-run 2025-09-07T07:35:59.9429131Z * [new branch] exclamaforte/gemm-export-model -> origin/exclamaforte/gemm-export-model 2025-09-07T07:35:59.9429630Z * [new branch] exclamaforte/gemm-model -> origin/exclamaforte/gemm-model 2025-09-07T07:35:59.9430298Z * [new branch] exclamaforte/gemm-model-all-data-collection -> origin/exclamaforte/gemm-model-all-data-collection 2025-09-07T07:35:59.9430742Z * [new branch] exclamaforte/gemm-to-amd -> origin/exclamaforte/gemm-to-amd 2025-09-07T07:35:59.9431251Z * [new branch] exclamaforte/just-gemm-model -> origin/exclamaforte/just-gemm-model 2025-09-07T07:35:59.9431847Z * [new branch] exclamaforte/just-gemm-model-no-refactor -> origin/exclamaforte/just-gemm-model-no-refactor 2025-09-07T07:35:59.9432287Z * [new branch] exclamaforte/max-autotune-ieee -> origin/exclamaforte/max-autotune-ieee 2025-09-07T07:35:59.9432863Z * [new branch] exclamaforte/memory-counter -> origin/exclamaforte/memory-counter 2025-09-07T07:35:59.9433461Z * [new branch] exclamaforte/profile-diff-algo -> origin/exclamaforte/profile-diff-algo 2025-09-07T07:35:59.9434070Z * [new branch] exclamaforte/profiler-combo -> origin/exclamaforte/profiler-combo 2025-09-07T07:35:59.9434670Z * [new branch] exclamaforte/test_cpp_wrapper_mode -> origin/exclamaforte/test_cpp_wrapper_mode 2025-09-07T07:35:59.9435194Z * [new branch] exclamaforte/update-autotune-configs -> origin/exclamaforte/update-autotune-configs 2025-09-07T07:35:59.9435745Z * [new branch] exclamaforte/update-autotune-configs-2 -> origin/exclamaforte/update-autotune-configs-2 2025-09-07T07:35:59.9436663Z * [new branch] exclamforte/gemm-model-final -> origin/exclamforte/gemm-model-final 2025-09-07T07:35:59.9437165Z * [new branch] exec -> origin/exec 2025-09-07T07:35:59.9437742Z * [new branch] executorch-module-shim -> origin/executorch-module-shim 2025-09-07T07:35:59.9438484Z * [new branch] experimental-mosaic -> origin/experimental-mosaic 2025-09-07T07:35:59.9438975Z * [new branch] export-D58091437 -> origin/export-D58091437 2025-09-07T07:35:59.9439714Z * [new branch] export-D61047529 -> origin/export-D61047529 2025-09-07T07:35:59.9440192Z * [new branch] export-D70112642 -> origin/export-D70112642 2025-09-07T07:35:59.9440904Z * [new branch] export-D71412006 -> origin/export-D71412006 2025-09-07T07:35:59.9441610Z * [new branch] export-D73042989 -> origin/export-D73042989 2025-09-07T07:35:59.9442169Z * [new branch] export-D75183591 -> origin/export-D75183591 2025-09-07T07:35:59.9442731Z * [new branch] export-D75617432 -> origin/export-D75617432 2025-09-07T07:35:59.9443812Z * [new branch] export-D75659965 -> origin/export-D75659965 2025-09-07T07:35:59.9444265Z * [new branch] export-D76080931 -> origin/export-D76080931 2025-09-07T07:35:59.9444802Z * [new branch] export-D76797250 -> origin/export-D76797250 2025-09-07T07:35:59.9445330Z * [new branch] export-D76885271 -> origin/export-D76885271 2025-09-07T07:35:59.9445873Z * [new branch] export-D76885620 -> origin/export-D76885620 2025-09-07T07:35:59.9446413Z * [new branch] export-D76936623 -> origin/export-D76936623 2025-09-07T07:35:59.9446977Z * [new branch] export-D76958268 -> origin/export-D76958268 2025-09-07T07:35:59.9447558Z * [new branch] export-D78375400 -> origin/export-D78375400 2025-09-07T07:35:59.9448561Z * [new branch] export-D78431305 -> origin/export-D78431305 2025-09-07T07:35:59.9449073Z * [new branch] export-D78580107 -> origin/export-D78580107 2025-09-07T07:35:59.9449612Z * [new branch] export-D78822171 -> origin/export-D78822171 2025-09-07T07:35:59.9450218Z * [new branch] export-D78822351 -> origin/export-D78822351 2025-09-07T07:35:59.9450669Z * [new branch] export-D78822507 -> origin/export-D78822507 2025-09-07T07:35:59.9451204Z * [new branch] export-D78826994 -> origin/export-D78826994 2025-09-07T07:35:59.9451746Z * [new branch] export-D78894324 -> origin/export-D78894324 2025-09-07T07:35:59.9452531Z * [new branch] export-D78929245 -> origin/export-D78929245 2025-09-07T07:35:59.9453010Z * [new branch] export-D78934925 -> origin/export-D78934925 2025-09-07T07:35:59.9453617Z * [new branch] export-D78953203 -> origin/export-D78953203 2025-09-07T07:35:59.9454171Z * [new branch] export-D78953229 -> origin/export-D78953229 2025-09-07T07:35:59.9454657Z * [new branch] export-D78957093 -> origin/export-D78957093 2025-09-07T07:35:59.9455147Z * [new branch] export-D78957389 -> origin/export-D78957389 2025-09-07T07:35:59.9455721Z * [new branch] export-D78996107 -> origin/export-D78996107 2025-09-07T07:35:59.9456249Z * [new branch] export-D79026433 -> origin/export-D79026433 2025-09-07T07:35:59.9456829Z * [new branch] export-D79230339 -> origin/export-D79230339 2025-09-07T07:35:59.9457414Z * [new branch] export-D79319835 -> origin/export-D79319835 2025-09-07T07:35:59.9457946Z * [new branch] export-D79328456 -> origin/export-D79328456 2025-09-07T07:35:59.9458531Z * [new branch] export-D79534608 -> origin/export-D79534608 2025-09-07T07:35:59.9459290Z * [new branch] export-D79785974 -> origin/export-D79785974 2025-09-07T07:35:59.9459839Z * [new branch] export-D80025417 -> origin/export-D80025417 2025-09-07T07:35:59.9460390Z * [new branch] export-D80120333 -> origin/export-D80120333 2025-09-07T07:35:59.9461244Z * [new branch] export-D80214882 -> origin/export-D80214882 2025-09-07T07:35:59.9461732Z * [new branch] export-D80319069 -> origin/export-D80319069 2025-09-07T07:35:59.9462292Z * [new branch] export-D80321215 -> origin/export-D80321215 2025-09-07T07:35:59.9462824Z * [new branch] export-D80503451 -> origin/export-D80503451 2025-09-07T07:35:59.9463334Z * [new branch] export-D80771648 -> origin/export-D80771648 2025-09-07T07:35:59.9463857Z * [new branch] export-D80823877 -> origin/export-D80823877 2025-09-07T07:35:59.9464410Z * [new branch] export-D80948073 -> origin/export-D80948073 2025-09-07T07:35:59.9465152Z * [new branch] export-D80958642 -> origin/export-D80958642 2025-09-07T07:35:59.9465685Z * [new branch] export-D80970483 -> origin/export-D80970483 2025-09-07T07:35:59.9466224Z * [new branch] export-D81054193 -> origin/export-D81054193 2025-09-07T07:35:59.9466776Z * [new branch] export-D81060182 -> origin/export-D81060182 2025-09-07T07:35:59.9467367Z * [new branch] export-D81078973 -> origin/export-D81078973 2025-09-07T07:35:59.9467934Z * [new branch] export-D81204584 -> origin/export-D81204584 2025-09-07T07:35:59.9468597Z * [new branch] export-D81284190 -> origin/export-D81284190 2025-09-07T07:35:59.9469354Z * [new branch] export-D81299840 -> origin/export-D81299840 2025-09-07T07:35:59.9469842Z * [new branch] export-D81429090 -> origin/export-D81429090 2025-09-07T07:35:59.9470375Z * [new branch] export-D81698719 -> origin/export-D81698719 2025-09-07T07:35:59.9470915Z * [new branch] export-D81747409 -> origin/export-D81747409 2025-09-07T07:35:59.9471783Z * [new branch] exported-model-train-idempotent -> origin/exported-model-train-idempotent 2025-09-07T07:35:59.9472581Z * [new branch] ezyang/wip-aot-descriptors -> origin/ezyang/wip-aot-descriptors 2025-09-07T07:35:59.9473002Z * [new branch] fa_u8_brgemm -> origin/fa_u8_brgemm 2025-09-07T07:35:59.9473614Z * [new branch] fastmath_baseline -> origin/fastmath_baseline 2025-09-07T07:35:59.9474603Z * [new branch] fbcode/warm -> origin/fbcode/warm 2025-09-07T07:35:59.9475192Z * [new branch] fca -> origin/fca 2025-09-07T07:35:59.9475738Z * [new branch] fca2_ca5984c -> origin/fca2_ca5984c 2025-09-07T07:35:59.9476277Z * [new branch] fca5 -> origin/fca5 2025-09-07T07:35:59.9477229Z * [new branch] feature/function-numa-binding -> origin/feature/function-numa-binding 2025-09-07T07:35:59.9477850Z * [new branch] feature/function-numa-binding-take2 -> origin/feature/function-numa-binding-take2 2025-09-07T07:35:59.9478252Z * [new branch] feature/numa-nproc-fix -> origin/feature/numa-nproc-fix 2025-09-07T07:35:59.9479037Z * [new branch] feature/numa-signpost-serialize -> origin/feature/numa-signpost-serialize 2025-09-07T07:35:59.9479497Z * [new branch] feature/parallel-numa-binding -> origin/feature/parallel-numa-binding 2025-09-07T07:35:59.9480394Z * [new branch] fengyuan/external-proj -> origin/fengyuan/external-proj 2025-09-07T07:35:59.9480913Z * [new branch] fengyuan/out-of-tree-xpu-ops-improve-test -> origin/fengyuan/out-of-tree-xpu-ops-improve-test 2025-09-07T07:35:59.9481450Z * [new branch] fengyuan/out-of-tree-xpu-ops-remove-dtype -> origin/fengyuan/out-of-tree-xpu-ops-remove-dtype 2025-09-07T07:35:59.9481860Z * [new branch] fengyuan/test-xpu -> origin/fengyuan/test-xpu 2025-09-07T07:35:59.9482941Z * [new branch] ffast_math_baseline -> origin/ffast_math_baseline 2025-09-07T07:35:59.9483405Z * [new branch] ffast_math_target -> origin/ffast_math_target 2025-09-07T07:35:59.9484345Z * [new branch] findhao/base_commit -> origin/findhao/base_commit 2025-09-07T07:35:59.9484823Z * [new branch] findhao/base_commit1 -> origin/findhao/base_commit1 2025-09-07T07:35:59.9485380Z * [new branch] findhao/multistream2 -> origin/findhao/multistream2 2025-09-07T07:35:59.9485941Z * [new branch] findhao/multistream5 -> origin/findhao/multistream5 2025-09-07T07:35:59.9486482Z * [new branch] findhao/multistream6 -> origin/findhao/multistream6 2025-09-07T07:35:59.9487028Z * [new branch] findhao/operatorbench3 -> origin/findhao/operatorbench3 2025-09-07T07:35:59.9487581Z * [new branch] findhao/operatorbench5 -> origin/findhao/operatorbench5 2025-09-07T07:35:59.9488316Z * [new branch] findhao/tritonparse -> origin/findhao/tritonparse 2025-09-07T07:35:59.9488797Z * [new branch] fix -> origin/fix 2025-09-07T07:35:59.9489433Z * [new branch] fix-ck-gemm-template-format -> origin/fix-ck-gemm-template-format 2025-09-07T07:35:59.9489988Z * [new branch] fix-config-ignore -> origin/fix-config-ignore 2025-09-07T07:35:59.9490549Z * [new branch] fix-dict-guard -> origin/fix-dict-guard 2025-09-07T07:35:59.9491137Z * [new branch] fix-inductor-periodic-0528 -> origin/fix-inductor-periodic-0528 2025-09-07T07:35:59.9491632Z * [new branch] fix-mps-benchmark -> origin/fix-mps-benchmark 2025-09-07T07:35:59.9492219Z * [new branch] fix-rlease-feature-template -> origin/fix-rlease-feature-template 2025-09-07T07:35:59.9492805Z * [new branch] fix-run-condition-upload-results -> origin/fix-run-condition-upload-results 2025-09-07T07:35:59.9493675Z * [new branch] fix-torchbench -> origin/fix-torchbench 2025-09-07T07:35:59.9494120Z * [new branch] fix_153389 -> origin/fix_153389 2025-09-07T07:35:59.9494722Z * [new branch] fix_fsdp_rs_bucket2 -> origin/fix_fsdp_rs_bucket2 2025-09-07T07:35:59.9495286Z * [new branch] fix_inductor_peridic_tests -> origin/fix_inductor_peridic_tests 2025-09-07T07:35:59.9495775Z * [new branch] fix_ubn_159469 -> origin/fix_ubn_159469 2025-09-07T07:35:59.9496358Z * [new branch] fixes-triage -> origin/fixes-triage 2025-09-07T07:35:59.9497088Z * [new branch] fixflashinfer -> origin/fixflashinfer 2025-09-07T07:35:59.9497563Z * [new branch] flash_decoding_cpu -> origin/flash_decoding_cpu 2025-09-07T07:35:59.9498510Z * [new branch] flex-flash -> origin/flex-flash 2025-09-07T07:35:59.9498936Z * [new branch] flex-lowering -> origin/flex-lowering 2025-09-07T07:35:59.9499490Z * [new branch] flex-warning -> origin/flex-warning 2025-09-07T07:35:59.9500132Z * [new branch] flex_attention_functorch_grad -> origin/flex_attention_functorch_grad 2025-09-07T07:35:59.9500981Z * [new branch] flex_flash -> origin/flex_flash 2025-09-07T07:35:59.9501550Z * [new branch] flexdecode-gqa-groups -> origin/flexdecode-gqa-groups 2025-09-07T07:35:59.9502500Z * [new branch] fmassa/fix_memeff_sharding_rule -> origin/fmassa/fix_memeff_sharding_rule 2025-09-07T07:35:59.9502969Z * [new branch] fsdp2_trace_rules -> origin/fsdp2_trace_rules 2025-09-07T07:35:59.9503551Z * [new branch] fsdpv2_3d -> origin/fsdpv2_3d 2025-09-07T07:35:59.9504257Z * [new branch] fsdpv2_3d_m1 -> origin/fsdpv2_3d_m1 2025-09-07T07:35:59.9504953Z * [new branch] fx_cpp -> origin/fx_cpp 2025-09-07T07:35:59.9505744Z * [new branch] fy/fix-win -> origin/fy/fix-win 2025-09-07T07:35:59.9507357Z * [new branch] gh/AlnisM/1/base -> origin/gh/AlnisM/1/base 2025-09-07T07:35:59.9507983Z * [new branch] gh/AlnisM/1/head -> origin/gh/AlnisM/1/head 2025-09-07T07:35:59.9509033Z * [new branch] gh/CaoE/2/base -> origin/gh/CaoE/2/base 2025-09-07T07:35:59.9509500Z * [new branch] gh/CaoE/2/head -> origin/gh/CaoE/2/head 2025-09-07T07:35:59.9510074Z * [new branch] gh/CaoE/2/orig -> origin/gh/CaoE/2/orig 2025-09-07T07:35:59.9511277Z * [new branch] gh/ColinPeppler/79/base -> origin/gh/ColinPeppler/79/base 2025-09-07T07:35:59.9511808Z * [new branch] gh/ColinPeppler/79/head -> origin/gh/ColinPeppler/79/head 2025-09-07T07:35:59.9512342Z * [new branch] gh/ColinPeppler/79/orig -> origin/gh/ColinPeppler/79/orig 2025-09-07T07:35:59.9513331Z * [new branch] gh/ColinPeppler/80/base -> origin/gh/ColinPeppler/80/base 2025-09-07T07:35:59.9513939Z * [new branch] gh/ColinPeppler/80/head -> origin/gh/ColinPeppler/80/head 2025-09-07T07:35:59.9514488Z * [new branch] gh/ColinPeppler/80/orig -> origin/gh/ColinPeppler/80/orig 2025-09-07T07:35:59.9515616Z * [new branch] gh/EikanWang/67/base -> origin/gh/EikanWang/67/base 2025-09-07T07:35:59.9516093Z * [new branch] gh/EikanWang/67/head -> origin/gh/EikanWang/67/head 2025-09-07T07:35:59.9516970Z * [new branch] gh/EikanWang/80/base -> origin/gh/EikanWang/80/base 2025-09-07T07:35:59.9517438Z * [new branch] gh/EikanWang/80/head -> origin/gh/EikanWang/80/head 2025-09-07T07:35:59.9518181Z * [new branch] gh/EikanWang/80/orig -> origin/gh/EikanWang/80/orig 2025-09-07T07:35:59.9518971Z * [new branch] gh/EikanWang/81/base -> origin/gh/EikanWang/81/base 2025-09-07T07:35:59.9519434Z * [new branch] gh/EikanWang/81/head -> origin/gh/EikanWang/81/head 2025-09-07T07:35:59.9519970Z * [new branch] gh/EikanWang/81/orig -> origin/gh/EikanWang/81/orig 2025-09-07T07:35:59.9520791Z * [new branch] gh/EikanWang/82/base -> origin/gh/EikanWang/82/base 2025-09-07T07:35:59.9521278Z * [new branch] gh/EikanWang/82/head -> origin/gh/EikanWang/82/head 2025-09-07T07:35:59.9521807Z * [new branch] gh/EikanWang/82/orig -> origin/gh/EikanWang/82/orig 2025-09-07T07:35:59.9523245Z * [new branch] gh/Gasoonjia/1/base -> origin/gh/Gasoonjia/1/base 2025-09-07T07:35:59.9523780Z * [new branch] gh/Gasoonjia/1/head -> origin/gh/Gasoonjia/1/head 2025-09-07T07:35:59.9524942Z * [new branch] gh/H-Huang/131/base -> origin/gh/H-Huang/131/base 2025-09-07T07:35:59.9525382Z * [new branch] gh/H-Huang/131/head -> origin/gh/H-Huang/131/head 2025-09-07T07:35:59.9525929Z * [new branch] gh/H-Huang/131/orig -> origin/gh/H-Huang/131/orig 2025-09-07T07:35:59.9526789Z * [new branch] gh/H-Huang/132/base -> origin/gh/H-Huang/132/base 2025-09-07T07:35:59.9527280Z * [new branch] gh/H-Huang/132/head -> origin/gh/H-Huang/132/head 2025-09-07T07:35:59.9527875Z * [new branch] gh/H-Huang/132/orig -> origin/gh/H-Huang/132/orig 2025-09-07T07:35:59.9528947Z * [new branch] gh/H-Huang/180/base -> origin/gh/H-Huang/180/base 2025-09-07T07:35:59.9529413Z * [new branch] gh/H-Huang/180/head -> origin/gh/H-Huang/180/head 2025-09-07T07:35:59.9529949Z * [new branch] gh/H-Huang/180/orig -> origin/gh/H-Huang/180/orig 2025-09-07T07:35:59.9530770Z * [new branch] gh/H-Huang/182/base -> origin/gh/H-Huang/182/base 2025-09-07T07:35:59.9531255Z * [new branch] gh/H-Huang/182/head -> origin/gh/H-Huang/182/head 2025-09-07T07:35:59.9531774Z * [new branch] gh/H-Huang/182/orig -> origin/gh/H-Huang/182/orig 2025-09-07T07:35:59.9532722Z * [new branch] gh/H-Huang/187/base -> origin/gh/H-Huang/187/base 2025-09-07T07:35:59.9533129Z * [new branch] gh/H-Huang/187/head -> origin/gh/H-Huang/187/head 2025-09-07T07:35:59.9533656Z * [new branch] gh/H-Huang/187/orig -> origin/gh/H-Huang/187/orig 2025-09-07T07:35:59.9534530Z * [new branch] gh/H-Huang/202/base -> origin/gh/H-Huang/202/base 2025-09-07T07:35:59.9534983Z * [new branch] gh/H-Huang/202/head -> origin/gh/H-Huang/202/head 2025-09-07T07:35:59.9535518Z * [new branch] gh/H-Huang/202/orig -> origin/gh/H-Huang/202/orig 2025-09-07T07:35:59.9536359Z * [new branch] gh/H-Huang/203/base -> origin/gh/H-Huang/203/base 2025-09-07T07:35:59.9536811Z * [new branch] gh/H-Huang/203/head -> origin/gh/H-Huang/203/head 2025-09-07T07:35:59.9537324Z * [new branch] gh/H-Huang/203/orig -> origin/gh/H-Huang/203/orig 2025-09-07T07:35:59.9538336Z * [new branch] gh/H-Huang/204/base -> origin/gh/H-Huang/204/base 2025-09-07T07:35:59.9538765Z * [new branch] gh/H-Huang/204/head -> origin/gh/H-Huang/204/head 2025-09-07T07:35:59.9539295Z * [new branch] gh/H-Huang/204/orig -> origin/gh/H-Huang/204/orig 2025-09-07T07:35:59.9540165Z * [new branch] gh/H-Huang/205/base -> origin/gh/H-Huang/205/base 2025-09-07T07:35:59.9540633Z * [new branch] gh/H-Huang/205/head -> origin/gh/H-Huang/205/head 2025-09-07T07:35:59.9541183Z * [new branch] gh/H-Huang/205/orig -> origin/gh/H-Huang/205/orig 2025-09-07T07:35:59.9542006Z * [new branch] gh/H-Huang/206/base -> origin/gh/H-Huang/206/base 2025-09-07T07:35:59.9542482Z * [new branch] gh/H-Huang/206/head -> origin/gh/H-Huang/206/head 2025-09-07T07:35:59.9543050Z * [new branch] gh/H-Huang/206/orig -> origin/gh/H-Huang/206/orig 2025-09-07T07:35:59.9543854Z * [new branch] gh/H-Huang/207/base -> origin/gh/H-Huang/207/base 2025-09-07T07:35:59.9544334Z * [new branch] gh/H-Huang/207/head -> origin/gh/H-Huang/207/head 2025-09-07T07:35:59.9544865Z * [new branch] gh/H-Huang/207/orig -> origin/gh/H-Huang/207/orig 2025-09-07T07:35:59.9545705Z * [new branch] gh/H-Huang/208/base -> origin/gh/H-Huang/208/base 2025-09-07T07:35:59.9546146Z * [new branch] gh/H-Huang/208/head -> origin/gh/H-Huang/208/head 2025-09-07T07:35:59.9546723Z * [new branch] gh/H-Huang/208/orig -> origin/gh/H-Huang/208/orig 2025-09-07T07:35:59.9547683Z * [new branch] gh/H-Huang/209/base -> origin/gh/H-Huang/209/base 2025-09-07T07:35:59.9548129Z * [new branch] gh/H-Huang/209/head -> origin/gh/H-Huang/209/head 2025-09-07T07:35:59.9548677Z * [new branch] gh/H-Huang/209/orig -> origin/gh/H-Huang/209/orig 2025-09-07T07:35:59.9549487Z * [new branch] gh/H-Huang/210/base -> origin/gh/H-Huang/210/base 2025-09-07T07:35:59.9549953Z * [new branch] gh/H-Huang/210/head -> origin/gh/H-Huang/210/head 2025-09-07T07:35:59.9550504Z * [new branch] gh/H-Huang/210/orig -> origin/gh/H-Huang/210/orig 2025-09-07T07:35:59.9551337Z * [new branch] gh/H-Huang/211/base -> origin/gh/H-Huang/211/base 2025-09-07T07:35:59.9551839Z * [new branch] gh/H-Huang/211/head -> origin/gh/H-Huang/211/head 2025-09-07T07:35:59.9552368Z * [new branch] gh/H-Huang/211/orig -> origin/gh/H-Huang/211/orig 2025-09-07T07:35:59.9553199Z * [new branch] gh/H-Huang/212/base -> origin/gh/H-Huang/212/base 2025-09-07T07:35:59.9553645Z * [new branch] gh/H-Huang/212/head -> origin/gh/H-Huang/212/head 2025-09-07T07:35:59.9554173Z * [new branch] gh/H-Huang/212/orig -> origin/gh/H-Huang/212/orig 2025-09-07T07:35:59.9555432Z * [new branch] gh/H-Huang/213/base -> origin/gh/H-Huang/213/base 2025-09-07T07:35:59.9555909Z * [new branch] gh/H-Huang/213/head -> origin/gh/H-Huang/213/head 2025-09-07T07:35:59.9556434Z * [new branch] gh/H-Huang/213/orig -> origin/gh/H-Huang/213/orig 2025-09-07T07:35:59.9557483Z * [new branch] gh/H-Huang/214/base -> origin/gh/H-Huang/214/base 2025-09-07T07:35:59.9558040Z * [new branch] gh/H-Huang/214/head -> origin/gh/H-Huang/214/head 2025-09-07T07:35:59.9558542Z * [new branch] gh/H-Huang/214/orig -> origin/gh/H-Huang/214/orig 2025-09-07T07:35:59.9559648Z * [new branch] gh/IvanKobzarev/112/base -> origin/gh/IvanKobzarev/112/base 2025-09-07T07:35:59.9560150Z * [new branch] gh/IvanKobzarev/112/head -> origin/gh/IvanKobzarev/112/head 2025-09-07T07:35:59.9560699Z * [new branch] gh/IvanKobzarev/112/orig -> origin/gh/IvanKobzarev/112/orig 2025-09-07T07:35:59.9561573Z * [new branch] gh/IvanKobzarev/115/base -> origin/gh/IvanKobzarev/115/base 2025-09-07T07:35:59.9562082Z * [new branch] gh/IvanKobzarev/115/head -> origin/gh/IvanKobzarev/115/head 2025-09-07T07:35:59.9562656Z * [new branch] gh/IvanKobzarev/115/orig -> origin/gh/IvanKobzarev/115/orig 2025-09-07T07:35:59.9563777Z * [new branch] gh/IvanKobzarev/116/base -> origin/gh/IvanKobzarev/116/base 2025-09-07T07:35:59.9564331Z * [new branch] gh/IvanKobzarev/116/head -> origin/gh/IvanKobzarev/116/head 2025-09-07T07:35:59.9564901Z * [new branch] gh/IvanKobzarev/116/orig -> origin/gh/IvanKobzarev/116/orig 2025-09-07T07:35:59.9565812Z * [new branch] gh/IvanKobzarev/118/base -> origin/gh/IvanKobzarev/118/base 2025-09-07T07:35:59.9566272Z * [new branch] gh/IvanKobzarev/118/head -> origin/gh/IvanKobzarev/118/head 2025-09-07T07:35:59.9566964Z * [new branch] gh/IvanKobzarev/118/orig -> origin/gh/IvanKobzarev/118/orig 2025-09-07T07:35:59.9567974Z * [new branch] gh/IvanKobzarev/126/base -> origin/gh/IvanKobzarev/126/base 2025-09-07T07:35:59.9568502Z * [new branch] gh/IvanKobzarev/126/head -> origin/gh/IvanKobzarev/126/head 2025-09-07T07:35:59.9569124Z * [new branch] gh/IvanKobzarev/126/orig -> origin/gh/IvanKobzarev/126/orig 2025-09-07T07:35:59.9570113Z * [new branch] gh/IvanKobzarev/127/base -> origin/gh/IvanKobzarev/127/base 2025-09-07T07:35:59.9570879Z * [new branch] gh/IvanKobzarev/127/head -> origin/gh/IvanKobzarev/127/head 2025-09-07T07:35:59.9571205Z * [new branch] gh/IvanKobzarev/127/orig -> origin/gh/IvanKobzarev/127/orig 2025-09-07T07:35:59.9572701Z * [new branch] gh/IvanKobzarev/128/base -> origin/gh/IvanKobzarev/128/base 2025-09-07T07:35:59.9573112Z * [new branch] gh/IvanKobzarev/128/head -> origin/gh/IvanKobzarev/128/head 2025-09-07T07:35:59.9573494Z * [new branch] gh/IvanKobzarev/128/orig -> origin/gh/IvanKobzarev/128/orig 2025-09-07T07:35:59.9573951Z * [new branch] gh/IvanKobzarev/132/base -> origin/gh/IvanKobzarev/132/base 2025-09-07T07:35:59.9574528Z * [new branch] gh/IvanKobzarev/132/head -> origin/gh/IvanKobzarev/132/head 2025-09-07T07:35:59.9575078Z * [new branch] gh/IvanKobzarev/132/orig -> origin/gh/IvanKobzarev/132/orig 2025-09-07T07:35:59.9576272Z * [new branch] gh/IvanKobzarev/133/base -> origin/gh/IvanKobzarev/133/base 2025-09-07T07:35:59.9576994Z * [new branch] gh/IvanKobzarev/133/head -> origin/gh/IvanKobzarev/133/head 2025-09-07T07:35:59.9577754Z * [new branch] gh/IvanKobzarev/133/orig -> origin/gh/IvanKobzarev/133/orig 2025-09-07T07:35:59.9578509Z * [new branch] gh/IvanKobzarev/134/base -> origin/gh/IvanKobzarev/134/base 2025-09-07T07:35:59.9578945Z * [new branch] gh/IvanKobzarev/134/head -> origin/gh/IvanKobzarev/134/head 2025-09-07T07:35:59.9579475Z * [new branch] gh/IvanKobzarev/134/orig -> origin/gh/IvanKobzarev/134/orig 2025-09-07T07:35:59.9580684Z * [new branch] gh/IvanKobzarev/135/base -> origin/gh/IvanKobzarev/135/base 2025-09-07T07:35:59.9581077Z * [new branch] gh/IvanKobzarev/135/head -> origin/gh/IvanKobzarev/135/head 2025-09-07T07:35:59.9581649Z * [new branch] gh/IvanKobzarev/135/orig -> origin/gh/IvanKobzarev/135/orig 2025-09-07T07:35:59.9582677Z * [new branch] gh/IvanKobzarev/136/base -> origin/gh/IvanKobzarev/136/base 2025-09-07T07:35:59.9583087Z * [new branch] gh/IvanKobzarev/136/head -> origin/gh/IvanKobzarev/136/head 2025-09-07T07:35:59.9583668Z * [new branch] gh/IvanKobzarev/136/orig -> origin/gh/IvanKobzarev/136/orig 2025-09-07T07:35:59.9584355Z * [new branch] gh/IvanKobzarev/137/base -> origin/gh/IvanKobzarev/137/base 2025-09-07T07:35:59.9584866Z * [new branch] gh/IvanKobzarev/137/head -> origin/gh/IvanKobzarev/137/head 2025-09-07T07:35:59.9585400Z * [new branch] gh/IvanKobzarev/137/orig -> origin/gh/IvanKobzarev/137/orig 2025-09-07T07:35:59.9586211Z * [new branch] gh/IvanKobzarev/138/base -> origin/gh/IvanKobzarev/138/base 2025-09-07T07:35:59.9586706Z * [new branch] gh/IvanKobzarev/138/head -> origin/gh/IvanKobzarev/138/head 2025-09-07T07:35:59.9587533Z * [new branch] gh/IvanKobzarev/138/orig -> origin/gh/IvanKobzarev/138/orig 2025-09-07T07:35:59.9588399Z * [new branch] gh/IvanKobzarev/139/base -> origin/gh/IvanKobzarev/139/base 2025-09-07T07:35:59.9588826Z * [new branch] gh/IvanKobzarev/139/head -> origin/gh/IvanKobzarev/139/head 2025-09-07T07:35:59.9589413Z * [new branch] gh/IvanKobzarev/139/orig -> origin/gh/IvanKobzarev/139/orig 2025-09-07T07:35:59.9590485Z * [new branch] gh/IvanKobzarev/140/base -> origin/gh/IvanKobzarev/140/base 2025-09-07T07:35:59.9590898Z * [new branch] gh/IvanKobzarev/140/head -> origin/gh/IvanKobzarev/140/head 2025-09-07T07:35:59.9591447Z * [new branch] gh/IvanKobzarev/140/orig -> origin/gh/IvanKobzarev/140/orig 2025-09-07T07:35:59.9592723Z * [new branch] gh/IvanKobzarev/141/base -> origin/gh/IvanKobzarev/141/base 2025-09-07T07:35:59.9593275Z * [new branch] gh/IvanKobzarev/141/head -> origin/gh/IvanKobzarev/141/head 2025-09-07T07:35:59.9593869Z * [new branch] gh/IvanKobzarev/141/orig -> origin/gh/IvanKobzarev/141/orig 2025-09-07T07:35:59.9594979Z * [new branch] gh/IvanKobzarev/142/base -> origin/gh/IvanKobzarev/142/base 2025-09-07T07:35:59.9595431Z * [new branch] gh/IvanKobzarev/142/head -> origin/gh/IvanKobzarev/142/head 2025-09-07T07:35:59.9597417Z * [new branch] gh/IvanKobzarev/142/orig -> origin/gh/IvanKobzarev/142/orig 2025-09-07T07:35:59.9597747Z * [new branch] gh/IvanKobzarev/143/base -> origin/gh/IvanKobzarev/143/base 2025-09-07T07:35:59.9598092Z * [new branch] gh/IvanKobzarev/143/head -> origin/gh/IvanKobzarev/143/head 2025-09-07T07:35:59.9598420Z * [new branch] gh/IvanKobzarev/143/orig -> origin/gh/IvanKobzarev/143/orig 2025-09-07T07:35:59.9599687Z * [new branch] gh/IvanKobzarev/144/base -> origin/gh/IvanKobzarev/144/base 2025-09-07T07:35:59.9600009Z * [new branch] gh/IvanKobzarev/144/head -> origin/gh/IvanKobzarev/144/head 2025-09-07T07:35:59.9600334Z * [new branch] gh/IvanKobzarev/144/orig -> origin/gh/IvanKobzarev/144/orig 2025-09-07T07:35:59.9600848Z * [new branch] gh/IvanKobzarev/145/base -> origin/gh/IvanKobzarev/145/base 2025-09-07T07:35:59.9601391Z * [new branch] gh/IvanKobzarev/145/head -> origin/gh/IvanKobzarev/145/head 2025-09-07T07:35:59.9601935Z * [new branch] gh/IvanKobzarev/145/orig -> origin/gh/IvanKobzarev/145/orig 2025-09-07T07:35:59.9602849Z * [new branch] gh/IvanKobzarev/146/base -> origin/gh/IvanKobzarev/146/base 2025-09-07T07:35:59.9603317Z * [new branch] gh/IvanKobzarev/146/head -> origin/gh/IvanKobzarev/146/head 2025-09-07T07:35:59.9603866Z * [new branch] gh/IvanKobzarev/146/orig -> origin/gh/IvanKobzarev/146/orig 2025-09-07T07:35:59.9605025Z * [new branch] gh/NikhilAPatel/1/base -> origin/gh/NikhilAPatel/1/base 2025-09-07T07:35:59.9605577Z * [new branch] gh/NikhilAPatel/1/head -> origin/gh/NikhilAPatel/1/head 2025-09-07T07:35:59.9606372Z * [new branch] gh/NikhilAPatel/2/base -> origin/gh/NikhilAPatel/2/base 2025-09-07T07:35:59.9606755Z * [new branch] gh/NikhilAPatel/2/head -> origin/gh/NikhilAPatel/2/head 2025-09-07T07:35:59.9607818Z * [new branch] gh/NikhilAPatel/4/base -> origin/gh/NikhilAPatel/4/base 2025-09-07T07:35:59.9608648Z * [new branch] gh/NikhilAPatel/4/head -> origin/gh/NikhilAPatel/4/head 2025-09-07T07:35:59.9609593Z * [new branch] gh/PaliC/1/base -> origin/gh/PaliC/1/base 2025-09-07T07:35:59.9610060Z * [new branch] gh/PaliC/1/head -> origin/gh/PaliC/1/head 2025-09-07T07:35:59.9610618Z * [new branch] gh/PaliC/1/orig -> origin/gh/PaliC/1/orig 2025-09-07T07:35:59.9611532Z * [new branch] gh/PaliC/17/base -> origin/gh/PaliC/17/base 2025-09-07T07:35:59.9611974Z * [new branch] gh/PaliC/17/head -> origin/gh/PaliC/17/head 2025-09-07T07:35:59.9612539Z * [new branch] gh/PaliC/17/orig -> origin/gh/PaliC/17/orig 2025-09-07T07:35:59.9613452Z * [new branch] gh/PaliC/18/base -> origin/gh/PaliC/18/base 2025-09-07T07:35:59.9613946Z * [new branch] gh/PaliC/18/head -> origin/gh/PaliC/18/head 2025-09-07T07:35:59.9614503Z * [new branch] gh/PaliC/18/orig -> origin/gh/PaliC/18/orig 2025-09-07T07:35:59.9615303Z * [new branch] gh/PaliC/2/base -> origin/gh/PaliC/2/base 2025-09-07T07:35:59.9615812Z * [new branch] gh/PaliC/2/head -> origin/gh/PaliC/2/head 2025-09-07T07:35:59.9616314Z * [new branch] gh/PaliC/2/orig -> origin/gh/PaliC/2/orig 2025-09-07T07:35:59.9617222Z * [new branch] gh/PaliC/20/base -> origin/gh/PaliC/20/base 2025-09-07T07:35:59.9617856Z * [new branch] gh/PaliC/20/head -> origin/gh/PaliC/20/head 2025-09-07T07:35:59.9618347Z * [new branch] gh/PaliC/20/orig -> origin/gh/PaliC/20/orig 2025-09-07T07:35:59.9619173Z * [new branch] gh/PaliC/21/base -> origin/gh/PaliC/21/base 2025-09-07T07:35:59.9619635Z * [new branch] gh/PaliC/21/head -> origin/gh/PaliC/21/head 2025-09-07T07:35:59.9620175Z * [new branch] gh/PaliC/21/orig -> origin/gh/PaliC/21/orig 2025-09-07T07:35:59.9620976Z * [new branch] gh/PaliC/22/base -> origin/gh/PaliC/22/base 2025-09-07T07:35:59.9621434Z * [new branch] gh/PaliC/22/head -> origin/gh/PaliC/22/head 2025-09-07T07:35:59.9621966Z * [new branch] gh/PaliC/22/orig -> origin/gh/PaliC/22/orig 2025-09-07T07:35:59.9622784Z * [new branch] gh/PaliC/23/base -> origin/gh/PaliC/23/base 2025-09-07T07:35:59.9623227Z * [new branch] gh/PaliC/23/head -> origin/gh/PaliC/23/head 2025-09-07T07:35:59.9623762Z * [new branch] gh/PaliC/23/orig -> origin/gh/PaliC/23/orig 2025-09-07T07:35:59.9624616Z * [new branch] gh/PaliC/24/base -> origin/gh/PaliC/24/base 2025-09-07T07:35:59.9625065Z * [new branch] gh/PaliC/24/head -> origin/gh/PaliC/24/head 2025-09-07T07:35:59.9625576Z * [new branch] gh/PaliC/24/orig -> origin/gh/PaliC/24/orig 2025-09-07T07:35:59.9626706Z * [new branch] gh/PaulZhang12/17/base -> origin/gh/PaulZhang12/17/base 2025-09-07T07:35:59.9627253Z * [new branch] gh/PaulZhang12/17/head -> origin/gh/PaulZhang12/17/head 2025-09-07T07:35:59.9628267Z * [new branch] gh/PaulZhang12/20/base -> origin/gh/PaulZhang12/20/base 2025-09-07T07:35:59.9628728Z * [new branch] gh/PaulZhang12/20/head -> origin/gh/PaulZhang12/20/head 2025-09-07T07:35:59.9629268Z * [new branch] gh/PaulZhang12/20/orig -> origin/gh/PaulZhang12/20/orig 2025-09-07T07:35:59.9630127Z * [new branch] gh/PaulZhang12/21/base -> origin/gh/PaulZhang12/21/base 2025-09-07T07:35:59.9630631Z * [new branch] gh/PaulZhang12/21/head -> origin/gh/PaulZhang12/21/head 2025-09-07T07:35:59.9631227Z * [new branch] gh/PaulZhang12/21/orig -> origin/gh/PaulZhang12/21/orig 2025-09-07T07:35:59.9632152Z * [new branch] gh/PaulZhang12/22/base -> origin/gh/PaulZhang12/22/base 2025-09-07T07:35:59.9632585Z * [new branch] gh/PaulZhang12/22/head -> origin/gh/PaulZhang12/22/head 2025-09-07T07:35:59.9633102Z * [new branch] gh/PaulZhang12/22/orig -> origin/gh/PaulZhang12/22/orig 2025-09-07T07:35:59.9634021Z * [new branch] gh/PaulZhang12/23/base -> origin/gh/PaulZhang12/23/base 2025-09-07T07:35:59.9634463Z * [new branch] gh/PaulZhang12/23/head -> origin/gh/PaulZhang12/23/head 2025-09-07T07:35:59.9634927Z * [new branch] gh/PaulZhang12/23/orig -> origin/gh/PaulZhang12/23/orig 2025-09-07T07:35:59.9635681Z * [new branch] gh/PaulZhang12/24/base -> origin/gh/PaulZhang12/24/base 2025-09-07T07:35:59.9636115Z * [new branch] gh/PaulZhang12/24/head -> origin/gh/PaulZhang12/24/head 2025-09-07T07:35:59.9636908Z * [new branch] gh/PaulZhang12/24/orig -> origin/gh/PaulZhang12/24/orig 2025-09-07T07:35:59.9637530Z * [new branch] gh/PaulZhang12/25/base -> origin/gh/PaulZhang12/25/base 2025-09-07T07:35:59.9638051Z * [new branch] gh/PaulZhang12/25/head -> origin/gh/PaulZhang12/25/head 2025-09-07T07:35:59.9638621Z * [new branch] gh/PaulZhang12/25/orig -> origin/gh/PaulZhang12/25/orig 2025-09-07T07:35:59.9639686Z * [new branch] gh/SamGinzburg/11/base -> origin/gh/SamGinzburg/11/base 2025-09-07T07:35:59.9640150Z * [new branch] gh/SamGinzburg/11/head -> origin/gh/SamGinzburg/11/head 2025-09-07T07:35:59.9641566Z * [new branch] gh/Sidharth123-cpu/24/base -> origin/gh/Sidharth123-cpu/24/base 2025-09-07T07:35:59.9642196Z * [new branch] gh/Sidharth123-cpu/25/base -> origin/gh/Sidharth123-cpu/25/base 2025-09-07T07:35:59.9642958Z * [new branch] gh/Sidharth123-cpu/26/base -> origin/gh/Sidharth123-cpu/26/base 2025-09-07T07:35:59.9643768Z * [new branch] gh/Sidharth123-cpu/27/base -> origin/gh/Sidharth123-cpu/27/base 2025-09-07T07:35:59.9644767Z * [new branch] gh/StrongerXi/1/base -> origin/gh/StrongerXi/1/base 2025-09-07T07:35:59.9645221Z * [new branch] gh/StrongerXi/1/head -> origin/gh/StrongerXi/1/head 2025-09-07T07:35:59.9646084Z * [new branch] gh/StrongerXi/133/base -> origin/gh/StrongerXi/133/base 2025-09-07T07:35:59.9646527Z * [new branch] gh/StrongerXi/133/head -> origin/gh/StrongerXi/133/head 2025-09-07T07:35:59.9647090Z * [new branch] gh/StrongerXi/133/orig -> origin/gh/StrongerXi/133/orig 2025-09-07T07:35:59.9648030Z * [new branch] gh/StrongerXi/134/base -> origin/gh/StrongerXi/134/base 2025-09-07T07:35:59.9648480Z * [new branch] gh/StrongerXi/134/head -> origin/gh/StrongerXi/134/head 2025-09-07T07:35:59.9649016Z * [new branch] gh/StrongerXi/134/orig -> origin/gh/StrongerXi/134/orig 2025-09-07T07:35:59.9650023Z * [new branch] gh/StrongerXi/136/base -> origin/gh/StrongerXi/136/base 2025-09-07T07:35:59.9650343Z * [new branch] gh/StrongerXi/136/head -> origin/gh/StrongerXi/136/head 2025-09-07T07:35:59.9650867Z * [new branch] gh/StrongerXi/136/orig -> origin/gh/StrongerXi/136/orig 2025-09-07T07:35:59.9651651Z * [new branch] gh/StrongerXi/137/base -> origin/gh/StrongerXi/137/base 2025-09-07T07:35:59.9652146Z * [new branch] gh/StrongerXi/137/head -> origin/gh/StrongerXi/137/head 2025-09-07T07:35:59.9652679Z * [new branch] gh/StrongerXi/137/orig -> origin/gh/StrongerXi/137/orig 2025-09-07T07:35:59.9653451Z * [new branch] gh/StrongerXi/138/base -> origin/gh/StrongerXi/138/base 2025-09-07T07:35:59.9653886Z * [new branch] gh/StrongerXi/138/head -> origin/gh/StrongerXi/138/head 2025-09-07T07:35:59.9654451Z * [new branch] gh/StrongerXi/138/orig -> origin/gh/StrongerXi/138/orig 2025-09-07T07:35:59.9655235Z * [new branch] gh/StrongerXi/139/base -> origin/gh/StrongerXi/139/base 2025-09-07T07:35:59.9655668Z * [new branch] gh/StrongerXi/139/head -> origin/gh/StrongerXi/139/head 2025-09-07T07:35:59.9656268Z * [new branch] gh/StrongerXi/139/orig -> origin/gh/StrongerXi/139/orig 2025-09-07T07:35:59.9657172Z * [new branch] gh/StrongerXi/140/base -> origin/gh/StrongerXi/140/base 2025-09-07T07:35:59.9657586Z * [new branch] gh/StrongerXi/140/head -> origin/gh/StrongerXi/140/head 2025-09-07T07:35:59.9658345Z * [new branch] gh/StrongerXi/140/orig -> origin/gh/StrongerXi/140/orig 2025-09-07T07:35:59.9659082Z * [new branch] gh/StrongerXi/71/base -> origin/gh/StrongerXi/71/base 2025-09-07T07:35:59.9659524Z * [new branch] gh/StrongerXi/71/head -> origin/gh/StrongerXi/71/head 2025-09-07T07:35:59.9660316Z * [new branch] gh/StrongerXi/72/base -> origin/gh/StrongerXi/72/base 2025-09-07T07:35:59.9660735Z * [new branch] gh/StrongerXi/72/head -> origin/gh/StrongerXi/72/head 2025-09-07T07:35:59.9661921Z * [new branch] gh/XilunWu/133/base -> origin/gh/XilunWu/133/base 2025-09-07T07:35:59.9662246Z * [new branch] gh/XilunWu/133/head -> origin/gh/XilunWu/133/head 2025-09-07T07:35:59.9662890Z * [new branch] gh/XilunWu/133/orig -> origin/gh/XilunWu/133/orig 2025-09-07T07:35:59.9663684Z * [new branch] gh/XilunWu/139/base -> origin/gh/XilunWu/139/base 2025-09-07T07:35:59.9664163Z * [new branch] gh/XilunWu/139/head -> origin/gh/XilunWu/139/head 2025-09-07T07:35:59.9664639Z * [new branch] gh/XilunWu/139/orig -> origin/gh/XilunWu/139/orig 2025-09-07T07:35:59.9665510Z * [new branch] gh/XilunWu/143/base -> origin/gh/XilunWu/143/base 2025-09-07T07:35:59.9666134Z * [new branch] gh/XilunWu/143/head -> origin/gh/XilunWu/143/head 2025-09-07T07:35:59.9666753Z * [new branch] gh/XilunWu/143/orig -> origin/gh/XilunWu/143/orig 2025-09-07T07:35:59.9667619Z * [new branch] gh/XilunWu/144/base -> origin/gh/XilunWu/144/base 2025-09-07T07:35:59.9668093Z * [new branch] gh/XilunWu/144/head -> origin/gh/XilunWu/144/head 2025-09-07T07:35:59.9668706Z * [new branch] gh/XilunWu/144/orig -> origin/gh/XilunWu/144/orig 2025-09-07T07:35:59.9669579Z * [new branch] gh/XilunWu/145/base -> origin/gh/XilunWu/145/base 2025-09-07T07:35:59.9669978Z * [new branch] gh/XilunWu/145/head -> origin/gh/XilunWu/145/head 2025-09-07T07:35:59.9670475Z * [new branch] gh/XilunWu/145/orig -> origin/gh/XilunWu/145/orig 2025-09-07T07:35:59.9671233Z * [new branch] gh/XilunWu/146/base -> origin/gh/XilunWu/146/base 2025-09-07T07:35:59.9671746Z * [new branch] gh/XilunWu/146/head -> origin/gh/XilunWu/146/head 2025-09-07T07:35:59.9672261Z * [new branch] gh/XilunWu/146/orig -> origin/gh/XilunWu/146/orig 2025-09-07T07:35:59.9673129Z * [new branch] gh/XilunWu/147/base -> origin/gh/XilunWu/147/base 2025-09-07T07:35:59.9673628Z * [new branch] gh/XilunWu/147/head -> origin/gh/XilunWu/147/head 2025-09-07T07:35:59.9674078Z * [new branch] gh/XilunWu/147/orig -> origin/gh/XilunWu/147/orig 2025-09-07T07:35:59.9674800Z * [new branch] gh/XilunWu/148/base -> origin/gh/XilunWu/148/base 2025-09-07T07:35:59.9675576Z * [new branch] gh/XilunWu/148/head -> origin/gh/XilunWu/148/head 2025-09-07T07:35:59.9676015Z * [new branch] gh/XilunWu/148/orig -> origin/gh/XilunWu/148/orig 2025-09-07T07:35:59.9676846Z * [new branch] gh/XilunWu/149/base -> origin/gh/XilunWu/149/base 2025-09-07T07:35:59.9677209Z * [new branch] gh/XilunWu/149/head -> origin/gh/XilunWu/149/head 2025-09-07T07:35:59.9677728Z * [new branch] gh/XilunWu/149/orig -> origin/gh/XilunWu/149/orig 2025-09-07T07:35:59.9678548Z * [new branch] gh/XilunWu/150/base -> origin/gh/XilunWu/150/base 2025-09-07T07:35:59.9678955Z * [new branch] gh/XilunWu/150/head -> origin/gh/XilunWu/150/head 2025-09-07T07:35:59.9679479Z * [new branch] gh/XilunWu/150/orig -> origin/gh/XilunWu/150/orig 2025-09-07T07:35:59.9680361Z * [new branch] gh/XilunWu/151/base -> origin/gh/XilunWu/151/base 2025-09-07T07:35:59.9680850Z * [new branch] gh/XilunWu/151/head -> origin/gh/XilunWu/151/head 2025-09-07T07:35:59.9681451Z * [new branch] gh/XilunWu/151/orig -> origin/gh/XilunWu/151/orig 2025-09-07T07:35:59.9682279Z * [new branch] gh/XilunWu/152/base -> origin/gh/XilunWu/152/base 2025-09-07T07:35:59.9682742Z * [new branch] gh/XilunWu/152/head -> origin/gh/XilunWu/152/head 2025-09-07T07:35:59.9683233Z * [new branch] gh/XilunWu/152/orig -> origin/gh/XilunWu/152/orig 2025-09-07T07:35:59.9684187Z * [new branch] gh/XilunWu/153/base -> origin/gh/XilunWu/153/base 2025-09-07T07:35:59.9685377Z * [new branch] gh/XilunWu/153/head -> origin/gh/XilunWu/153/head 2025-09-07T07:35:59.9685769Z * [new branch] gh/XilunWu/153/orig -> origin/gh/XilunWu/153/orig 2025-09-07T07:35:59.9686267Z * [new branch] gh/XilunWu/160/base -> origin/gh/XilunWu/160/base 2025-09-07T07:35:59.9686744Z * [new branch] gh/XilunWu/160/head -> origin/gh/XilunWu/160/head 2025-09-07T07:35:59.9687323Z * [new branch] gh/XilunWu/160/orig -> origin/gh/XilunWu/160/orig 2025-09-07T07:35:59.9688277Z * [new branch] gh/XilunWu/161/base -> origin/gh/XilunWu/161/base 2025-09-07T07:35:59.9688713Z * [new branch] gh/XilunWu/161/head -> origin/gh/XilunWu/161/head 2025-09-07T07:35:59.9689240Z * [new branch] gh/XilunWu/161/orig -> origin/gh/XilunWu/161/orig 2025-09-07T07:35:59.9690145Z * [new branch] gh/XilunWu/163/base -> origin/gh/XilunWu/163/base 2025-09-07T07:35:59.9690597Z * [new branch] gh/XilunWu/163/head -> origin/gh/XilunWu/163/head 2025-09-07T07:35:59.9691188Z * [new branch] gh/XilunWu/163/orig -> origin/gh/XilunWu/163/orig 2025-09-07T07:35:59.9692353Z * [new branch] gh/XilunWu/164/base -> origin/gh/XilunWu/164/base 2025-09-07T07:35:59.9692881Z * [new branch] gh/XilunWu/164/head -> origin/gh/XilunWu/164/head 2025-09-07T07:35:59.9693413Z * [new branch] gh/XilunWu/164/orig -> origin/gh/XilunWu/164/orig 2025-09-07T07:35:59.9694373Z * [new branch] gh/XilunWu/165/base -> origin/gh/XilunWu/165/base 2025-09-07T07:35:59.9695137Z * [new branch] gh/XilunWu/165/head -> origin/gh/XilunWu/165/head 2025-09-07T07:35:59.9695646Z * [new branch] gh/XilunWu/165/orig -> origin/gh/XilunWu/165/orig 2025-09-07T07:35:59.9696624Z * [new branch] gh/XilunWu/166/base -> origin/gh/XilunWu/166/base 2025-09-07T07:35:59.9697149Z * [new branch] gh/XilunWu/166/head -> origin/gh/XilunWu/166/head 2025-09-07T07:35:59.9697664Z * [new branch] gh/XilunWu/166/orig -> origin/gh/XilunWu/166/orig 2025-09-07T07:35:59.9698512Z * [new branch] gh/XilunWu/167/base -> origin/gh/XilunWu/167/base 2025-09-07T07:35:59.9698996Z * [new branch] gh/XilunWu/167/head -> origin/gh/XilunWu/167/head 2025-09-07T07:35:59.9699545Z * [new branch] gh/XilunWu/167/orig -> origin/gh/XilunWu/167/orig 2025-09-07T07:35:59.9700458Z * [new branch] gh/XilunWu/168/base -> origin/gh/XilunWu/168/base 2025-09-07T07:35:59.9700862Z * [new branch] gh/XilunWu/168/head -> origin/gh/XilunWu/168/head 2025-09-07T07:35:59.9701368Z * [new branch] gh/XilunWu/168/orig -> origin/gh/XilunWu/168/orig 2025-09-07T07:35:59.9702217Z * [new branch] gh/XilunWu/169/base -> origin/gh/XilunWu/169/base 2025-09-07T07:35:59.9702691Z * [new branch] gh/XilunWu/169/head -> origin/gh/XilunWu/169/head 2025-09-07T07:35:59.9703198Z * [new branch] gh/XilunWu/169/orig -> origin/gh/XilunWu/169/orig 2025-09-07T07:35:59.9703970Z * [new branch] gh/XilunWu/170/base -> origin/gh/XilunWu/170/base 2025-09-07T07:35:59.9704703Z * [new branch] gh/XilunWu/170/head -> origin/gh/XilunWu/170/head 2025-09-07T07:35:59.9705124Z * [new branch] gh/XilunWu/170/orig -> origin/gh/XilunWu/170/orig 2025-09-07T07:35:59.9706272Z * [new branch] gh/XuehaiPan/14/base -> origin/gh/XuehaiPan/14/base 2025-09-07T07:35:59.9706789Z * [new branch] gh/XuehaiPan/14/head -> origin/gh/XuehaiPan/14/head 2025-09-07T07:35:59.9707256Z * [new branch] gh/XuehaiPan/14/orig -> origin/gh/XuehaiPan/14/orig 2025-09-07T07:35:59.9708088Z * [new branch] gh/XuehaiPan/179/base -> origin/gh/XuehaiPan/179/base 2025-09-07T07:35:59.9708604Z * [new branch] gh/XuehaiPan/179/head -> origin/gh/XuehaiPan/179/head 2025-09-07T07:35:59.9709193Z * [new branch] gh/XuehaiPan/179/orig -> origin/gh/XuehaiPan/179/orig 2025-09-07T07:35:59.9710127Z * [new branch] gh/XuehaiPan/189/base -> origin/gh/XuehaiPan/189/base 2025-09-07T07:35:59.9710606Z * [new branch] gh/XuehaiPan/189/head -> origin/gh/XuehaiPan/189/head 2025-09-07T07:35:59.9711140Z * [new branch] gh/XuehaiPan/189/orig -> origin/gh/XuehaiPan/189/orig 2025-09-07T07:35:59.9712031Z * [new branch] gh/XuehaiPan/232/base -> origin/gh/XuehaiPan/232/base 2025-09-07T07:35:59.9712524Z * [new branch] gh/XuehaiPan/232/head -> origin/gh/XuehaiPan/232/head 2025-09-07T07:35:59.9713068Z * [new branch] gh/XuehaiPan/232/orig -> origin/gh/XuehaiPan/232/orig 2025-09-07T07:35:59.9714044Z * [new branch] gh/XuehaiPan/249/base -> origin/gh/XuehaiPan/249/base 2025-09-07T07:35:59.9714587Z * [new branch] gh/XuehaiPan/249/head -> origin/gh/XuehaiPan/249/head 2025-09-07T07:35:59.9715125Z * [new branch] gh/XuehaiPan/249/orig -> origin/gh/XuehaiPan/249/orig 2025-09-07T07:35:59.9715856Z * [new branch] gh/XuehaiPan/253/base -> origin/gh/XuehaiPan/253/base 2025-09-07T07:35:59.9716337Z * [new branch] gh/XuehaiPan/253/head -> origin/gh/XuehaiPan/253/head 2025-09-07T07:35:59.9716881Z * [new branch] gh/XuehaiPan/253/orig -> origin/gh/XuehaiPan/253/orig 2025-09-07T07:35:59.9717626Z * [new branch] gh/XuehaiPan/254/base -> origin/gh/XuehaiPan/254/base 2025-09-07T07:35:59.9718088Z * [new branch] gh/XuehaiPan/254/head -> origin/gh/XuehaiPan/254/head 2025-09-07T07:35:59.9718686Z * [new branch] gh/XuehaiPan/254/orig -> origin/gh/XuehaiPan/254/orig 2025-09-07T07:35:59.9719512Z * [new branch] gh/XuehaiPan/255/base -> origin/gh/XuehaiPan/255/base 2025-09-07T07:35:59.9719973Z * [new branch] gh/XuehaiPan/255/head -> origin/gh/XuehaiPan/255/head 2025-09-07T07:35:59.9720463Z * [new branch] gh/XuehaiPan/255/orig -> origin/gh/XuehaiPan/255/orig 2025-09-07T07:35:59.9721296Z * [new branch] gh/XuehaiPan/257/base -> origin/gh/XuehaiPan/257/base 2025-09-07T07:35:59.9721738Z * [new branch] gh/XuehaiPan/257/head -> origin/gh/XuehaiPan/257/head 2025-09-07T07:35:59.9722291Z * [new branch] gh/XuehaiPan/257/orig -> origin/gh/XuehaiPan/257/orig 2025-09-07T07:35:59.9723231Z * [new branch] gh/XuehaiPan/271/base -> origin/gh/XuehaiPan/271/base 2025-09-07T07:35:59.9723698Z * [new branch] gh/XuehaiPan/271/head -> origin/gh/XuehaiPan/271/head 2025-09-07T07:35:59.9724230Z * [new branch] gh/XuehaiPan/271/orig -> origin/gh/XuehaiPan/271/orig 2025-09-07T07:35:59.9725069Z * [new branch] gh/XuehaiPan/290/base -> origin/gh/XuehaiPan/290/base 2025-09-07T07:35:59.9725585Z * [new branch] gh/XuehaiPan/290/head -> origin/gh/XuehaiPan/290/head 2025-09-07T07:35:59.9726055Z * [new branch] gh/XuehaiPan/290/orig -> origin/gh/XuehaiPan/290/orig 2025-09-07T07:35:59.9726957Z * [new branch] gh/XuehaiPan/343/base -> origin/gh/XuehaiPan/343/base 2025-09-07T07:35:59.9727425Z * [new branch] gh/XuehaiPan/343/head -> origin/gh/XuehaiPan/343/head 2025-09-07T07:35:59.9727964Z * [new branch] gh/XuehaiPan/343/orig -> origin/gh/XuehaiPan/343/orig 2025-09-07T07:35:59.9728821Z * [new branch] gh/XuehaiPan/347/base -> origin/gh/XuehaiPan/347/base 2025-09-07T07:35:59.9729295Z * [new branch] gh/XuehaiPan/347/head -> origin/gh/XuehaiPan/347/head 2025-09-07T07:35:59.9729880Z * [new branch] gh/XuehaiPan/347/orig -> origin/gh/XuehaiPan/347/orig 2025-09-07T07:35:59.9739562Z * [new branch] gh/XuehaiPan/348/base -> origin/gh/XuehaiPan/348/base 2025-09-07T07:35:59.9739871Z * [new branch] gh/XuehaiPan/348/head -> origin/gh/XuehaiPan/348/head 2025-09-07T07:35:59.9740174Z * [new branch] gh/XuehaiPan/348/orig -> origin/gh/XuehaiPan/348/orig 2025-09-07T07:35:59.9740500Z * [new branch] gh/XuehaiPan/350/base -> origin/gh/XuehaiPan/350/base 2025-09-07T07:35:59.9740806Z * [new branch] gh/XuehaiPan/350/head -> origin/gh/XuehaiPan/350/head 2025-09-07T07:35:59.9741115Z * [new branch] gh/XuehaiPan/350/orig -> origin/gh/XuehaiPan/350/orig 2025-09-07T07:35:59.9741421Z * [new branch] gh/XuehaiPan/356/base -> origin/gh/XuehaiPan/356/base 2025-09-07T07:35:59.9741721Z * [new branch] gh/XuehaiPan/356/head -> origin/gh/XuehaiPan/356/head 2025-09-07T07:35:59.9742028Z * [new branch] gh/XuehaiPan/356/orig -> origin/gh/XuehaiPan/356/orig 2025-09-07T07:35:59.9742329Z * [new branch] gh/XuehaiPan/357/base -> origin/gh/XuehaiPan/357/base 2025-09-07T07:35:59.9742632Z * [new branch] gh/XuehaiPan/357/head -> origin/gh/XuehaiPan/357/head 2025-09-07T07:35:59.9742933Z * [new branch] gh/XuehaiPan/357/orig -> origin/gh/XuehaiPan/357/orig 2025-09-07T07:35:59.9743230Z * [new branch] gh/XuehaiPan/358/base -> origin/gh/XuehaiPan/358/base 2025-09-07T07:35:59.9743531Z * [new branch] gh/XuehaiPan/358/head -> origin/gh/XuehaiPan/358/head 2025-09-07T07:35:59.9743833Z * [new branch] gh/XuehaiPan/358/orig -> origin/gh/XuehaiPan/358/orig 2025-09-07T07:35:59.9744133Z * [new branch] gh/XuehaiPan/359/base -> origin/gh/XuehaiPan/359/base 2025-09-07T07:35:59.9744554Z * [new branch] gh/XuehaiPan/359/head -> origin/gh/XuehaiPan/359/head 2025-09-07T07:35:59.9744863Z * [new branch] gh/XuehaiPan/359/orig -> origin/gh/XuehaiPan/359/orig 2025-09-07T07:35:59.9745167Z * [new branch] gh/XuehaiPan/360/base -> origin/gh/XuehaiPan/360/base 2025-09-07T07:35:59.9745472Z * [new branch] gh/XuehaiPan/360/head -> origin/gh/XuehaiPan/360/head 2025-09-07T07:35:59.9745778Z * [new branch] gh/XuehaiPan/360/orig -> origin/gh/XuehaiPan/360/orig 2025-09-07T07:35:59.9746082Z * [new branch] gh/XuehaiPan/365/base -> origin/gh/XuehaiPan/365/base 2025-09-07T07:35:59.9746380Z * [new branch] gh/XuehaiPan/365/head -> origin/gh/XuehaiPan/365/head 2025-09-07T07:35:59.9746687Z * [new branch] gh/XuehaiPan/365/orig -> origin/gh/XuehaiPan/365/orig 2025-09-07T07:35:59.9746997Z * [new branch] gh/XuehaiPan/366/base -> origin/gh/XuehaiPan/366/base 2025-09-07T07:35:59.9747305Z * [new branch] gh/XuehaiPan/366/head -> origin/gh/XuehaiPan/366/head 2025-09-07T07:35:59.9747608Z * [new branch] gh/XuehaiPan/369/base -> origin/gh/XuehaiPan/369/base 2025-09-07T07:35:59.9747910Z * [new branch] gh/XuehaiPan/369/head -> origin/gh/XuehaiPan/369/head 2025-09-07T07:35:59.9748214Z * [new branch] gh/XuehaiPan/369/orig -> origin/gh/XuehaiPan/369/orig 2025-09-07T07:35:59.9748733Z * [new branch] gh/XuehaiPan/370/base -> origin/gh/XuehaiPan/370/base 2025-09-07T07:35:59.9749268Z * [new branch] gh/XuehaiPan/370/head -> origin/gh/XuehaiPan/370/head 2025-09-07T07:35:59.9749836Z * [new branch] gh/XuehaiPan/370/orig -> origin/gh/XuehaiPan/370/orig 2025-09-07T07:35:59.9750676Z * [new branch] gh/XuehaiPan/380/base -> origin/gh/XuehaiPan/380/base 2025-09-07T07:35:59.9751409Z * [new branch] gh/XuehaiPan/380/head -> origin/gh/XuehaiPan/380/head 2025-09-07T07:35:59.9751911Z * [new branch] gh/XuehaiPan/380/orig -> origin/gh/XuehaiPan/380/orig 2025-09-07T07:35:59.9752756Z * [new branch] gh/XuehaiPan/381/base -> origin/gh/XuehaiPan/381/base 2025-09-07T07:35:59.9753199Z * [new branch] gh/XuehaiPan/381/head -> origin/gh/XuehaiPan/381/head 2025-09-07T07:35:59.9754080Z * [new branch] gh/XuehaiPan/382/base -> origin/gh/XuehaiPan/382/base 2025-09-07T07:35:59.9754536Z * [new branch] gh/XuehaiPan/382/head -> origin/gh/XuehaiPan/382/head 2025-09-07T07:35:59.9755100Z * [new branch] gh/XuehaiPan/382/orig -> origin/gh/XuehaiPan/382/orig 2025-09-07T07:35:59.9755996Z * [new branch] gh/XuehaiPan/383/base -> origin/gh/XuehaiPan/383/base 2025-09-07T07:35:59.9756430Z * [new branch] gh/XuehaiPan/383/head -> origin/gh/XuehaiPan/383/head 2025-09-07T07:35:59.9756978Z * [new branch] gh/XuehaiPan/383/orig -> origin/gh/XuehaiPan/383/orig 2025-09-07T07:35:59.9757818Z * [new branch] gh/XuehaiPan/384/base -> origin/gh/XuehaiPan/384/base 2025-09-07T07:35:59.9758723Z * [new branch] gh/XuehaiPan/384/head -> origin/gh/XuehaiPan/384/head 2025-09-07T07:35:59.9759134Z * [new branch] gh/XuehaiPan/384/orig -> origin/gh/XuehaiPan/384/orig 2025-09-07T07:35:59.9760017Z * [new branch] gh/XuehaiPan/385/base -> origin/gh/XuehaiPan/385/base 2025-09-07T07:35:59.9760606Z * [new branch] gh/XuehaiPan/385/head -> origin/gh/XuehaiPan/385/head 2025-09-07T07:35:59.9761078Z * [new branch] gh/XuehaiPan/385/orig -> origin/gh/XuehaiPan/385/orig 2025-09-07T07:35:59.9761877Z * [new branch] gh/XuehaiPan/386/base -> origin/gh/XuehaiPan/386/base 2025-09-07T07:35:59.9762327Z * [new branch] gh/XuehaiPan/386/head -> origin/gh/XuehaiPan/386/head 2025-09-07T07:35:59.9762885Z * [new branch] gh/XuehaiPan/386/orig -> origin/gh/XuehaiPan/386/orig 2025-09-07T07:35:59.9763707Z * [new branch] gh/XuehaiPan/387/base -> origin/gh/XuehaiPan/387/base 2025-09-07T07:35:59.9764146Z * [new branch] gh/XuehaiPan/387/head -> origin/gh/XuehaiPan/387/head 2025-09-07T07:35:59.9764687Z * [new branch] gh/XuehaiPan/387/orig -> origin/gh/XuehaiPan/387/orig 2025-09-07T07:35:59.9765737Z * [new branch] gh/ZainRizvi/1/base -> origin/gh/ZainRizvi/1/base 2025-09-07T07:35:59.9766197Z * [new branch] gh/ZainRizvi/1/head -> origin/gh/ZainRizvi/1/head 2025-09-07T07:35:59.9766957Z * [new branch] gh/ZainRizvi/2/base -> origin/gh/ZainRizvi/2/base 2025-09-07T07:35:59.9767356Z * [new branch] gh/ZainRizvi/2/head -> origin/gh/ZainRizvi/2/head 2025-09-07T07:35:59.9768137Z * [new branch] gh/ZainRizvi/3/base -> origin/gh/ZainRizvi/3/base 2025-09-07T07:35:59.9768634Z * [new branch] gh/ZainRizvi/3/head -> origin/gh/ZainRizvi/3/head 2025-09-07T07:35:59.9769488Z * [new branch] gh/ZainRizvi/4/base -> origin/gh/ZainRizvi/4/base 2025-09-07T07:35:59.9769964Z * [new branch] gh/ZainRizvi/4/head -> origin/gh/ZainRizvi/4/head 2025-09-07T07:35:59.9770797Z * [new branch] gh/ZainRizvi/5/base -> origin/gh/ZainRizvi/5/base 2025-09-07T07:35:59.9771270Z * [new branch] gh/ZainRizvi/5/head -> origin/gh/ZainRizvi/5/head 2025-09-07T07:35:59.9772056Z * [new branch] gh/ZainRizvi/6/base -> origin/gh/ZainRizvi/6/base 2025-09-07T07:35:59.9772516Z * [new branch] gh/ZainRizvi/6/head -> origin/gh/ZainRizvi/6/head 2025-09-07T07:35:59.9773035Z * [new branch] gh/ZainRizvi/6/orig -> origin/gh/ZainRizvi/6/orig 2025-09-07T07:35:59.9773858Z * [new branch] gh/ZainRizvi/7/base -> origin/gh/ZainRizvi/7/base 2025-09-07T07:35:59.9774396Z * [new branch] gh/ZainRizvi/7/head -> origin/gh/ZainRizvi/7/head 2025-09-07T07:35:59.9774836Z * [new branch] gh/ZainRizvi/7/orig -> origin/gh/ZainRizvi/7/orig 2025-09-07T07:35:59.9775725Z * [new branch] gh/ZainRizvi/8/base -> origin/gh/ZainRizvi/8/base 2025-09-07T07:35:59.9776213Z * [new branch] gh/ZainRizvi/8/head -> origin/gh/ZainRizvi/8/head 2025-09-07T07:35:59.9777030Z * [new branch] gh/ZainRizvi/9/base -> origin/gh/ZainRizvi/9/base 2025-09-07T07:35:59.9777491Z * [new branch] gh/ZainRizvi/9/head -> origin/gh/ZainRizvi/9/head 2025-09-07T07:35:59.9778030Z * [new branch] gh/ZainRizvi/9/orig -> origin/gh/ZainRizvi/9/orig 2025-09-07T07:35:59.9779063Z * [new branch] gh/ZhiweiYan-96/39/base -> origin/gh/ZhiweiYan-96/39/base 2025-09-07T07:35:59.9779866Z * [new branch] gh/ZhiweiYan-96/39/head -> origin/gh/ZhiweiYan-96/39/head 2025-09-07T07:35:59.9780354Z * [new branch] gh/ZhiweiYan-96/39/orig -> origin/gh/ZhiweiYan-96/39/orig 2025-09-07T07:35:59.9781188Z * [new branch] gh/ZhiweiYan-96/44/base -> origin/gh/ZhiweiYan-96/44/base 2025-09-07T07:35:59.9781644Z * [new branch] gh/ZhiweiYan-96/44/head -> origin/gh/ZhiweiYan-96/44/head 2025-09-07T07:35:59.9782424Z * [new branch] gh/ZhiweiYan-96/45/base -> origin/gh/ZhiweiYan-96/45/base 2025-09-07T07:35:59.9782831Z * [new branch] gh/ZhiweiYan-96/45/head -> origin/gh/ZhiweiYan-96/45/head 2025-09-07T07:35:59.9783758Z * [new branch] gh/ZhiweiYan-96/49/base -> origin/gh/ZhiweiYan-96/49/base 2025-09-07T07:35:59.9784209Z * [new branch] gh/ZhiweiYan-96/49/head -> origin/gh/ZhiweiYan-96/49/head 2025-09-07T07:35:59.9785018Z * [new branch] gh/ZhiweiYan-96/62/base -> origin/gh/ZhiweiYan-96/62/base 2025-09-07T07:35:59.9785484Z * [new branch] gh/ZhiweiYan-96/62/head -> origin/gh/ZhiweiYan-96/62/head 2025-09-07T07:35:59.9786313Z * [new branch] gh/ZhiweiYan-96/64/base -> origin/gh/ZhiweiYan-96/64/base 2025-09-07T07:35:59.9786776Z * [new branch] gh/ZhiweiYan-96/64/head -> origin/gh/ZhiweiYan-96/64/head 2025-09-07T07:35:59.9787382Z * [new branch] gh/ZhiweiYan-96/64/orig -> origin/gh/ZhiweiYan-96/64/orig 2025-09-07T07:35:59.9788191Z * [new branch] gh/ZhiweiYan-96/65/base -> origin/gh/ZhiweiYan-96/65/base 2025-09-07T07:35:59.9788641Z * [new branch] gh/ZhiweiYan-96/65/head -> origin/gh/ZhiweiYan-96/65/head 2025-09-07T07:35:59.9789382Z * [new branch] gh/ZhiweiYan-96/65/orig -> origin/gh/ZhiweiYan-96/65/orig 2025-09-07T07:35:59.9790151Z * [new branch] gh/ZhiweiYan-96/66/base -> origin/gh/ZhiweiYan-96/66/base 2025-09-07T07:35:59.9790616Z * [new branch] gh/ZhiweiYan-96/66/head -> origin/gh/ZhiweiYan-96/66/head 2025-09-07T07:35:59.9791412Z * [new branch] gh/ZhiweiYan-96/67/base -> origin/gh/ZhiweiYan-96/67/base 2025-09-07T07:35:59.9791840Z * [new branch] gh/ZhiweiYan-96/67/head -> origin/gh/ZhiweiYan-96/67/head 2025-09-07T07:35:59.9792593Z * [new branch] gh/ZhiweiYan-96/68/base -> origin/gh/ZhiweiYan-96/68/base 2025-09-07T07:35:59.9793018Z * [new branch] gh/ZhiweiYan-96/68/head -> origin/gh/ZhiweiYan-96/68/head 2025-09-07T07:35:59.9793581Z * [new branch] gh/ZhiweiYan-96/68/orig -> origin/gh/ZhiweiYan-96/68/orig 2025-09-07T07:35:59.9794652Z * [new branch] gh/aakhundov/1/base -> origin/gh/aakhundov/1/base 2025-09-07T07:35:59.9795137Z * [new branch] gh/aakhundov/1/head -> origin/gh/aakhundov/1/head 2025-09-07T07:35:59.9795860Z * [new branch] gh/aakhundov/2/base -> origin/gh/aakhundov/2/base 2025-09-07T07:35:59.9796269Z * [new branch] gh/aakhundov/2/head -> origin/gh/aakhundov/2/head 2025-09-07T07:35:59.9797164Z * [new branch] gh/aditew01/openblas -> origin/gh/aditew01/openblas 2025-09-07T07:35:59.9797594Z * [new branch] gh/aditew01/sbgemm -> origin/gh/aditew01/sbgemm 2025-09-07T07:35:59.9798124Z * [new branch] gh/aditew01/vecbf16 -> origin/gh/aditew01/vecbf16 2025-09-07T07:35:59.9799227Z * [new branch] gh/alexbrauckmann/paddedtensor_faketensor_init -> origin/gh/alexbrauckmann/paddedtensor_faketensor_init 2025-09-07T07:35:59.9799968Z * [new branch] gh/alexsamardzic/9/base -> origin/gh/alexsamardzic/9/base 2025-09-07T07:35:59.9800512Z * [new branch] gh/alexsamardzic/9/head -> origin/gh/alexsamardzic/9/head 2025-09-07T07:35:59.9801264Z * [new branch] gh/alexsamardzic/9/orig -> origin/gh/alexsamardzic/9/orig 2025-09-07T07:35:59.9802206Z * [new branch] gh/amjames/18/base -> origin/gh/amjames/18/base 2025-09-07T07:35:59.9802651Z * [new branch] gh/amjames/18/head -> origin/gh/amjames/18/head 2025-09-07T07:35:59.9803149Z * [new branch] gh/amjames/18/orig -> origin/gh/amjames/18/orig 2025-09-07T07:35:59.9804293Z * [new branch] gh/andrewor14/35/base -> origin/gh/andrewor14/35/base 2025-09-07T07:35:59.9804829Z * [new branch] gh/andrewor14/35/head -> origin/gh/andrewor14/35/head 2025-09-07T07:35:59.9805379Z * [new branch] gh/andrewor14/35/orig -> origin/gh/andrewor14/35/orig 2025-09-07T07:35:59.9806374Z * [new branch] gh/andrewor14/50/base -> origin/gh/andrewor14/50/base 2025-09-07T07:35:59.9806914Z * [new branch] gh/andrewor14/50/head -> origin/gh/andrewor14/50/head 2025-09-07T07:35:59.9807444Z * [new branch] gh/andrewor14/50/orig -> origin/gh/andrewor14/50/orig 2025-09-07T07:35:59.9808315Z * [new branch] gh/andrewor14/51/base -> origin/gh/andrewor14/51/base 2025-09-07T07:35:59.9809049Z * [new branch] gh/andrewor14/51/orig -> origin/gh/andrewor14/51/orig 2025-09-07T07:35:59.9810115Z * [new branch] gh/andyanwang/1/base -> origin/gh/andyanwang/1/base 2025-09-07T07:35:59.9810543Z * [new branch] gh/andyanwang/1/head -> origin/gh/andyanwang/1/head 2025-09-07T07:35:59.9811067Z * [new branch] gh/andyanwang/1/orig -> origin/gh/andyanwang/1/orig 2025-09-07T07:35:59.9812020Z * [new branch] gh/andyanwang/13/base -> origin/gh/andyanwang/13/base 2025-09-07T07:35:59.9812593Z * [new branch] gh/andyanwang/13/head -> origin/gh/andyanwang/13/head 2025-09-07T07:35:59.9813789Z * [new branch] gh/andyanwang/13/orig -> origin/gh/andyanwang/13/orig 2025-09-07T07:35:59.9814583Z * [new branch] gh/andyanwang/2/base -> origin/gh/andyanwang/2/base 2025-09-07T07:35:59.9815083Z * [new branch] gh/andyanwang/2/head -> origin/gh/andyanwang/2/head 2025-09-07T07:35:59.9815642Z * [new branch] gh/andyanwang/2/orig -> origin/gh/andyanwang/2/orig 2025-09-07T07:35:59.9816600Z * [new branch] gh/andyanwang/28/base -> origin/gh/andyanwang/28/base 2025-09-07T07:35:59.9817130Z * [new branch] gh/andyanwang/28/head -> origin/gh/andyanwang/28/head 2025-09-07T07:35:59.9818059Z * [new branch] gh/andyanwang/28/orig -> origin/gh/andyanwang/28/orig 2025-09-07T07:35:59.9818643Z * [new branch] gh/andyanwang/3/base -> origin/gh/andyanwang/3/base 2025-09-07T07:35:59.9819231Z * [new branch] gh/andyanwang/3/head -> origin/gh/andyanwang/3/head 2025-09-07T07:35:59.9820016Z * [new branch] gh/andyanwang/3/orig -> origin/gh/andyanwang/3/orig 2025-09-07T07:35:59.9820773Z * [new branch] gh/andyanwang/30/base -> origin/gh/andyanwang/30/base 2025-09-07T07:35:59.9821444Z * [new branch] gh/andyanwang/30/orig -> origin/gh/andyanwang/30/orig 2025-09-07T07:35:59.9822259Z * [new branch] gh/andyanwang/31/base -> origin/gh/andyanwang/31/base 2025-09-07T07:35:59.9823019Z * [new branch] gh/andyanwang/31/orig -> origin/gh/andyanwang/31/orig 2025-09-07T07:35:59.9824059Z * [new branch] gh/andyanwang/32/base -> origin/gh/andyanwang/32/base 2025-09-07T07:35:59.9824524Z * [new branch] gh/andyanwang/32/head -> origin/gh/andyanwang/32/head 2025-09-07T07:35:59.9825313Z * [new branch] gh/andyanwang/32/orig -> origin/gh/andyanwang/32/orig 2025-09-07T07:35:59.9826094Z * [new branch] gh/andyanwang/39/base -> origin/gh/andyanwang/39/base 2025-09-07T07:35:59.9826615Z * [new branch] gh/andyanwang/39/head -> origin/gh/andyanwang/39/head 2025-09-07T07:35:59.9827145Z * [new branch] gh/andyanwang/39/orig -> origin/gh/andyanwang/39/orig 2025-09-07T07:35:59.9828071Z * [new branch] gh/andyanwang/4/base -> origin/gh/andyanwang/4/base 2025-09-07T07:35:59.9828474Z * [new branch] gh/andyanwang/4/head -> origin/gh/andyanwang/4/head 2025-09-07T07:35:59.9829066Z * [new branch] gh/andyanwang/4/orig -> origin/gh/andyanwang/4/orig 2025-09-07T07:35:59.9830083Z * [new branch] gh/angelayi/107/base -> origin/gh/angelayi/107/base 2025-09-07T07:35:59.9830702Z * [new branch] gh/angelayi/107/head -> origin/gh/angelayi/107/head 2025-09-07T07:35:59.9831519Z * [new branch] gh/angelayi/111/base -> origin/gh/angelayi/111/base 2025-09-07T07:35:59.9831940Z * [new branch] gh/angelayi/111/head -> origin/gh/angelayi/111/head 2025-09-07T07:35:59.9832562Z * [new branch] gh/angelayi/111/orig -> origin/gh/angelayi/111/orig 2025-09-07T07:35:59.9833384Z * [new branch] gh/angelayi/112/base -> origin/gh/angelayi/112/base 2025-09-07T07:35:59.9833907Z * [new branch] gh/angelayi/112/head -> origin/gh/angelayi/112/head 2025-09-07T07:35:59.9834636Z * [new branch] gh/angelayi/112/orig -> origin/gh/angelayi/112/orig 2025-09-07T07:35:59.9835553Z * [new branch] gh/angelayi/113/base -> origin/gh/angelayi/113/base 2025-09-07T07:35:59.9835949Z * [new branch] gh/angelayi/113/head -> origin/gh/angelayi/113/head 2025-09-07T07:35:59.9836442Z * [new branch] gh/angelayi/113/orig -> origin/gh/angelayi/113/orig 2025-09-07T07:35:59.9837303Z * [new branch] gh/angelayi/114/base -> origin/gh/angelayi/114/base 2025-09-07T07:35:59.9837727Z * [new branch] gh/angelayi/114/head -> origin/gh/angelayi/114/head 2025-09-07T07:35:59.9838256Z * [new branch] gh/angelayi/114/orig -> origin/gh/angelayi/114/orig 2025-09-07T07:35:59.9839110Z * [new branch] gh/angelayi/115/base -> origin/gh/angelayi/115/base 2025-09-07T07:35:59.9839597Z * [new branch] gh/angelayi/115/head -> origin/gh/angelayi/115/head 2025-09-07T07:35:59.9840413Z * [new branch] gh/angelayi/115/orig -> origin/gh/angelayi/115/orig 2025-09-07T07:35:59.9841470Z * [new branch] gh/anijain2305/753/base -> origin/gh/anijain2305/753/base 2025-09-07T07:35:59.9841942Z * [new branch] gh/anijain2305/753/head -> origin/gh/anijain2305/753/head 2025-09-07T07:35:59.9842468Z * [new branch] gh/anijain2305/753/orig -> origin/gh/anijain2305/753/orig 2025-09-07T07:35:59.9843359Z * [new branch] gh/anijain2305/766/base -> origin/gh/anijain2305/766/base 2025-09-07T07:35:59.9843751Z * [new branch] gh/anijain2305/766/head -> origin/gh/anijain2305/766/head 2025-09-07T07:35:59.9844272Z * [new branch] gh/anijain2305/766/orig -> origin/gh/anijain2305/766/orig 2025-09-07T07:35:59.9845186Z * [new branch] gh/anijain2305/790/base -> origin/gh/anijain2305/790/base 2025-09-07T07:35:59.9845635Z * [new branch] gh/anijain2305/790/head -> origin/gh/anijain2305/790/head 2025-09-07T07:35:59.9846149Z * [new branch] gh/anijain2305/790/orig -> origin/gh/anijain2305/790/orig 2025-09-07T07:35:59.9846979Z * [new branch] gh/anijain2305/792/base -> origin/gh/anijain2305/792/base 2025-09-07T07:35:59.9847597Z * [new branch] gh/anijain2305/792/head -> origin/gh/anijain2305/792/head 2025-09-07T07:35:59.9848047Z * [new branch] gh/anijain2305/792/orig -> origin/gh/anijain2305/792/orig 2025-09-07T07:35:59.9848900Z * [new branch] gh/anijain2305/803/base -> origin/gh/anijain2305/803/base 2025-09-07T07:35:59.9849377Z * [new branch] gh/anijain2305/803/head -> origin/gh/anijain2305/803/head 2025-09-07T07:35:59.9850151Z * [new branch] gh/anijain2305/803/orig -> origin/gh/anijain2305/803/orig 2025-09-07T07:35:59.9850920Z * [new branch] gh/anijain2305/804/base -> origin/gh/anijain2305/804/base 2025-09-07T07:35:59.9851351Z * [new branch] gh/anijain2305/804/head -> origin/gh/anijain2305/804/head 2025-09-07T07:35:59.9851898Z * [new branch] gh/anijain2305/804/orig -> origin/gh/anijain2305/804/orig 2025-09-07T07:35:59.9852730Z * [new branch] gh/anijain2305/805/base -> origin/gh/anijain2305/805/base 2025-09-07T07:35:59.9853190Z * [new branch] gh/anijain2305/805/head -> origin/gh/anijain2305/805/head 2025-09-07T07:35:59.9853712Z * [new branch] gh/anijain2305/805/orig -> origin/gh/anijain2305/805/orig 2025-09-07T07:35:59.9854704Z * [new branch] gh/anijain2305/810/base -> origin/gh/anijain2305/810/base 2025-09-07T07:35:59.9855241Z * [new branch] gh/anijain2305/810/head -> origin/gh/anijain2305/810/head 2025-09-07T07:35:59.9855791Z * [new branch] gh/anijain2305/810/orig -> origin/gh/anijain2305/810/orig 2025-09-07T07:35:59.9856650Z * [new branch] gh/anijain2305/812/base -> origin/gh/anijain2305/812/base 2025-09-07T07:35:59.9857156Z * [new branch] gh/anijain2305/812/head -> origin/gh/anijain2305/812/head 2025-09-07T07:35:59.9857713Z * [new branch] gh/anijain2305/812/orig -> origin/gh/anijain2305/812/orig 2025-09-07T07:35:59.9858521Z * [new branch] gh/anijain2305/838/base -> origin/gh/anijain2305/838/base 2025-09-07T07:35:59.9859152Z * [new branch] gh/anijain2305/838/head -> origin/gh/anijain2305/838/head 2025-09-07T07:35:59.9859704Z * [new branch] gh/anijain2305/838/orig -> origin/gh/anijain2305/838/orig 2025-09-07T07:35:59.9860479Z * [new branch] gh/anijain2305/839/base -> origin/gh/anijain2305/839/base 2025-09-07T07:35:59.9860913Z * [new branch] gh/anijain2305/839/head -> origin/gh/anijain2305/839/head 2025-09-07T07:35:59.9861434Z * [new branch] gh/anijain2305/839/orig -> origin/gh/anijain2305/839/orig 2025-09-07T07:35:59.9862228Z * [new branch] gh/anijain2305/843/base -> origin/gh/anijain2305/843/base 2025-09-07T07:35:59.9862737Z * [new branch] gh/anijain2305/843/head -> origin/gh/anijain2305/843/head 2025-09-07T07:35:59.9863267Z * [new branch] gh/anijain2305/843/orig -> origin/gh/anijain2305/843/orig 2025-09-07T07:35:59.9864147Z * [new branch] gh/anijain2305/844/base -> origin/gh/anijain2305/844/base 2025-09-07T07:35:59.9864571Z * [new branch] gh/anijain2305/844/head -> origin/gh/anijain2305/844/head 2025-09-07T07:35:59.9865088Z * [new branch] gh/anijain2305/844/orig -> origin/gh/anijain2305/844/orig 2025-09-07T07:35:59.9865952Z * [new branch] gh/anijain2305/846/base -> origin/gh/anijain2305/846/base 2025-09-07T07:35:59.9866419Z * [new branch] gh/anijain2305/846/head -> origin/gh/anijain2305/846/head 2025-09-07T07:35:59.9866916Z * [new branch] gh/anijain2305/846/orig -> origin/gh/anijain2305/846/orig 2025-09-07T07:35:59.9867767Z * [new branch] gh/anijain2305/848/base -> origin/gh/anijain2305/848/base 2025-09-07T07:35:59.9868555Z * [new branch] gh/anijain2305/848/head -> origin/gh/anijain2305/848/head 2025-09-07T07:35:59.9870557Z * [new branch] gh/anijain2305/848/orig -> origin/gh/anijain2305/848/orig 2025-09-07T07:35:59.9871347Z * [new branch] gh/anijain2305/849/base -> origin/gh/anijain2305/849/base 2025-09-07T07:35:59.9871893Z * [new branch] gh/anijain2305/849/head -> origin/gh/anijain2305/849/head 2025-09-07T07:35:59.9872402Z * [new branch] gh/anijain2305/849/orig -> origin/gh/anijain2305/849/orig 2025-09-07T07:35:59.9873489Z * [new branch] gh/anijain2305/850/base -> origin/gh/anijain2305/850/base 2025-09-07T07:35:59.9873939Z * [new branch] gh/anijain2305/850/head -> origin/gh/anijain2305/850/head 2025-09-07T07:35:59.9874493Z * [new branch] gh/anijain2305/850/orig -> origin/gh/anijain2305/850/orig 2025-09-07T07:35:59.9875372Z * [new branch] gh/anijain2305/851/base -> origin/gh/anijain2305/851/base 2025-09-07T07:35:59.9875904Z * [new branch] gh/anijain2305/851/head -> origin/gh/anijain2305/851/head 2025-09-07T07:35:59.9876403Z * [new branch] gh/anijain2305/851/orig -> origin/gh/anijain2305/851/orig 2025-09-07T07:35:59.9877362Z * [new branch] gh/anijain2305/852/base -> origin/gh/anijain2305/852/base 2025-09-07T07:35:59.9877830Z * [new branch] gh/anijain2305/852/head -> origin/gh/anijain2305/852/head 2025-09-07T07:35:59.9878366Z * [new branch] gh/anijain2305/852/orig -> origin/gh/anijain2305/852/orig 2025-09-07T07:35:59.9879222Z * [new branch] gh/anijain2305/853/base -> origin/gh/anijain2305/853/base 2025-09-07T07:35:59.9879997Z * [new branch] gh/anijain2305/853/head -> origin/gh/anijain2305/853/head 2025-09-07T07:35:59.9880437Z * [new branch] gh/anijain2305/853/orig -> origin/gh/anijain2305/853/orig 2025-09-07T07:35:59.9881296Z * [new branch] gh/anijain2305/854/base -> origin/gh/anijain2305/854/base 2025-09-07T07:35:59.9881793Z * [new branch] gh/anijain2305/854/head -> origin/gh/anijain2305/854/head 2025-09-07T07:35:59.9882325Z * [new branch] gh/anijain2305/854/orig -> origin/gh/anijain2305/854/orig 2025-09-07T07:35:59.9883154Z * [new branch] gh/anijain2305/855/base -> origin/gh/anijain2305/855/base 2025-09-07T07:35:59.9883658Z * [new branch] gh/anijain2305/855/head -> origin/gh/anijain2305/855/head 2025-09-07T07:35:59.9884198Z * [new branch] gh/anijain2305/855/orig -> origin/gh/anijain2305/855/orig 2025-09-07T07:35:59.9885064Z * [new branch] gh/anijain2305/856/base -> origin/gh/anijain2305/856/base 2025-09-07T07:35:59.9885512Z * [new branch] gh/anijain2305/856/head -> origin/gh/anijain2305/856/head 2025-09-07T07:35:59.9886041Z * [new branch] gh/anijain2305/856/orig -> origin/gh/anijain2305/856/orig 2025-09-07T07:35:59.9886870Z * [new branch] gh/anijain2305/857/base -> origin/gh/anijain2305/857/base 2025-09-07T07:35:59.9887348Z * [new branch] gh/anijain2305/857/head -> origin/gh/anijain2305/857/head 2025-09-07T07:35:59.9887855Z * [new branch] gh/anijain2305/857/orig -> origin/gh/anijain2305/857/orig 2025-09-07T07:35:59.9888701Z * [new branch] gh/anijain2305/858/base -> origin/gh/anijain2305/858/base 2025-09-07T07:35:59.9889577Z * [new branch] gh/anijain2305/858/head -> origin/gh/anijain2305/858/head 2025-09-07T07:35:59.9889945Z * [new branch] gh/anijain2305/858/orig -> origin/gh/anijain2305/858/orig 2025-09-07T07:35:59.9890757Z * [new branch] gh/anijain2305/859/base -> origin/gh/anijain2305/859/base 2025-09-07T07:35:59.9891238Z * [new branch] gh/anijain2305/859/head -> origin/gh/anijain2305/859/head 2025-09-07T07:35:59.9891746Z * [new branch] gh/anijain2305/859/orig -> origin/gh/anijain2305/859/orig 2025-09-07T07:35:59.9892569Z * [new branch] gh/anijain2305/860/base -> origin/gh/anijain2305/860/base 2025-09-07T07:35:59.9893043Z * [new branch] gh/anijain2305/860/head -> origin/gh/anijain2305/860/head 2025-09-07T07:35:59.9893572Z * [new branch] gh/anijain2305/860/orig -> origin/gh/anijain2305/860/orig 2025-09-07T07:35:59.9894423Z * [new branch] gh/anijain2305/861/base -> origin/gh/anijain2305/861/base 2025-09-07T07:35:59.9895284Z * [new branch] gh/anijain2305/861/head -> origin/gh/anijain2305/861/head 2025-09-07T07:35:59.9895596Z * [new branch] gh/anijain2305/861/orig -> origin/gh/anijain2305/861/orig 2025-09-07T07:35:59.9896629Z * [new branch] gh/anijain2305/862/base -> origin/gh/anijain2305/862/base 2025-09-07T07:35:59.9896950Z * [new branch] gh/anijain2305/862/head -> origin/gh/anijain2305/862/head 2025-09-07T07:35:59.9897495Z * [new branch] gh/anijain2305/862/orig -> origin/gh/anijain2305/862/orig 2025-09-07T07:35:59.9898376Z * [new branch] gh/anijain2305/863/base -> origin/gh/anijain2305/863/base 2025-09-07T07:35:59.9899160Z * [new branch] gh/anijain2305/863/head -> origin/gh/anijain2305/863/head 2025-09-07T07:35:59.9899591Z * [new branch] gh/anijain2305/863/orig -> origin/gh/anijain2305/863/orig 2025-09-07T07:35:59.9900485Z * [new branch] gh/anijain2305/864/base -> origin/gh/anijain2305/864/base 2025-09-07T07:35:59.9900952Z * [new branch] gh/anijain2305/864/head -> origin/gh/anijain2305/864/head 2025-09-07T07:35:59.9901465Z * [new branch] gh/anijain2305/864/orig -> origin/gh/anijain2305/864/orig 2025-09-07T07:35:59.9902292Z * [new branch] gh/anijain2305/865/base -> origin/gh/anijain2305/865/base 2025-09-07T07:35:59.9902810Z * [new branch] gh/anijain2305/865/head -> origin/gh/anijain2305/865/head 2025-09-07T07:35:59.9903359Z * [new branch] gh/anijain2305/865/orig -> origin/gh/anijain2305/865/orig 2025-09-07T07:35:59.9904239Z * [new branch] gh/anijain2305/866/base -> origin/gh/anijain2305/866/base 2025-09-07T07:35:59.9904691Z * [new branch] gh/anijain2305/866/head -> origin/gh/anijain2305/866/head 2025-09-07T07:35:59.9905216Z * [new branch] gh/anijain2305/866/orig -> origin/gh/anijain2305/866/orig 2025-09-07T07:35:59.9906340Z * [new branch] gh/anjali411/216/base -> origin/gh/anjali411/216/base 2025-09-07T07:35:59.9906781Z * [new branch] gh/anjali411/216/head -> origin/gh/anjali411/216/head 2025-09-07T07:35:59.9907322Z * [new branch] gh/anjali411/216/orig -> origin/gh/anjali411/216/orig 2025-09-07T07:35:59.9908511Z * [new branch] gh/ankitageorge/13/base -> origin/gh/ankitageorge/13/base 2025-09-07T07:35:59.9908978Z * [new branch] gh/ankitageorge/13/head -> origin/gh/ankitageorge/13/head 2025-09-07T07:35:59.9909561Z * [new branch] gh/ankitageorge/13/orig -> origin/gh/ankitageorge/13/orig 2025-09-07T07:35:59.9910609Z * [new branch] gh/ankitageorge/14/base -> origin/gh/ankitageorge/14/base 2025-09-07T07:35:59.9911031Z * [new branch] gh/ankitageorge/14/head -> origin/gh/ankitageorge/14/head 2025-09-07T07:35:59.9911770Z * [new branch] gh/ankitageorge/14/orig -> origin/gh/ankitageorge/14/orig 2025-09-07T07:35:59.9912620Z * [new branch] gh/ankitageorge/15/base -> origin/gh/ankitageorge/15/base 2025-09-07T07:35:59.9913096Z * [new branch] gh/ankitageorge/15/head -> origin/gh/ankitageorge/15/head 2025-09-07T07:35:59.9913683Z * [new branch] gh/ankitageorge/15/orig -> origin/gh/ankitageorge/15/orig 2025-09-07T07:35:59.9914574Z * [new branch] gh/ankitageorge/16/base -> origin/gh/ankitageorge/16/base 2025-09-07T07:35:59.9915140Z * [new branch] gh/ankitageorge/16/head -> origin/gh/ankitageorge/16/head 2025-09-07T07:35:59.9915739Z * [new branch] gh/ankitageorge/16/orig -> origin/gh/ankitageorge/16/orig 2025-09-07T07:35:59.9916761Z * [new branch] gh/ankitageorge/17/base -> origin/gh/ankitageorge/17/base 2025-09-07T07:35:59.9917179Z * [new branch] gh/ankitageorge/17/head -> origin/gh/ankitageorge/17/head 2025-09-07T07:35:59.9917724Z * [new branch] gh/ankitageorge/17/orig -> origin/gh/ankitageorge/17/orig 2025-09-07T07:35:59.9918841Z * [new branch] gh/ankitageorge/21/base -> origin/gh/ankitageorge/21/base 2025-09-07T07:35:59.9919277Z * [new branch] gh/ankitageorge/21/head -> origin/gh/ankitageorge/21/head 2025-09-07T07:35:59.9919830Z * [new branch] gh/ankitageorge/21/orig -> origin/gh/ankitageorge/21/orig 2025-09-07T07:35:59.9920957Z * [new branch] gh/anshul-si/1/base -> origin/gh/anshul-si/1/base 2025-09-07T07:35:59.9921419Z * [new branch] gh/anshul-si/1/head -> origin/gh/anshul-si/1/head 2025-09-07T07:35:59.9922245Z * [new branch] gh/anshul-si/15/base -> origin/gh/anshul-si/15/base 2025-09-07T07:35:59.9922720Z * [new branch] gh/anshul-si/15/head -> origin/gh/anshul-si/15/head 2025-09-07T07:35:59.9923263Z * [new branch] gh/anshul-si/15/orig -> origin/gh/anshul-si/15/orig 2025-09-07T07:35:59.9924220Z * [new branch] gh/anshul-si/16/base -> origin/gh/anshul-si/16/base 2025-09-07T07:35:59.9924649Z * [new branch] gh/anshul-si/16/head -> origin/gh/anshul-si/16/head 2025-09-07T07:35:59.9925176Z * [new branch] gh/anshul-si/16/orig -> origin/gh/anshul-si/16/orig 2025-09-07T07:35:59.9926195Z * [new branch] gh/anshul-si/17/base -> origin/gh/anshul-si/17/base 2025-09-07T07:35:59.9926715Z * [new branch] gh/anshul-si/17/head -> origin/gh/anshul-si/17/head 2025-09-07T07:35:59.9927447Z * [new branch] gh/anshul-si/17/orig -> origin/gh/anshul-si/17/orig 2025-09-07T07:35:59.9928336Z * [new branch] gh/anshul-si/18/base -> origin/gh/anshul-si/18/base 2025-09-07T07:35:59.9928927Z * [new branch] gh/anshul-si/18/head -> origin/gh/anshul-si/18/head 2025-09-07T07:35:59.9929502Z * [new branch] gh/anshul-si/18/orig -> origin/gh/anshul-si/18/orig 2025-09-07T07:35:59.9930375Z * [new branch] gh/anshul-si/19/base -> origin/gh/anshul-si/19/base 2025-09-07T07:35:59.9930905Z * [new branch] gh/anshul-si/19/head -> origin/gh/anshul-si/19/head 2025-09-07T07:35:59.9931446Z * [new branch] gh/anshul-si/19/orig -> origin/gh/anshul-si/19/orig 2025-09-07T07:35:59.9932158Z * [new branch] gh/anshul-si/2/base -> origin/gh/anshul-si/2/base 2025-09-07T07:35:59.9932589Z * [new branch] gh/anshul-si/2/head -> origin/gh/anshul-si/2/head 2025-09-07T07:35:59.9933676Z * [new branch] gh/anshul-si/20/base -> origin/gh/anshul-si/20/base 2025-09-07T07:35:59.9934166Z * [new branch] gh/anshul-si/20/head -> origin/gh/anshul-si/20/head 2025-09-07T07:35:59.9934739Z * [new branch] gh/anshul-si/20/orig -> origin/gh/anshul-si/20/orig 2025-09-07T07:35:59.9935525Z * [new branch] gh/anshul-si/21/base -> origin/gh/anshul-si/21/base 2025-09-07T07:35:59.9936222Z * [new branch] gh/anshul-si/21/head -> origin/gh/anshul-si/21/head 2025-09-07T07:35:59.9936538Z * [new branch] gh/anshul-si/21/orig -> origin/gh/anshul-si/21/orig 2025-09-07T07:35:59.9937594Z * [new branch] gh/anshul-si/22/base -> origin/gh/anshul-si/22/base 2025-09-07T07:35:59.9937967Z * [new branch] gh/anshul-si/22/head -> origin/gh/anshul-si/22/head 2025-09-07T07:35:59.9938718Z * [new branch] gh/anshul-si/22/orig -> origin/gh/anshul-si/22/orig 2025-09-07T07:35:59.9939354Z * [new branch] gh/anshul-si/23/base -> origin/gh/anshul-si/23/base 2025-09-07T07:35:59.9939888Z * [new branch] gh/anshul-si/23/head -> origin/gh/anshul-si/23/head 2025-09-07T07:35:59.9940424Z * [new branch] gh/anshul-si/23/orig -> origin/gh/anshul-si/23/orig 2025-09-07T07:35:59.9941349Z * [new branch] gh/anshul-si/24/base -> origin/gh/anshul-si/24/base 2025-09-07T07:35:59.9942097Z * [new branch] gh/anshul-si/24/head -> origin/gh/anshul-si/24/head 2025-09-07T07:35:59.9942407Z * [new branch] gh/anshul-si/24/orig -> origin/gh/anshul-si/24/orig 2025-09-07T07:35:59.9943307Z * [new branch] gh/anshul-si/25/base -> origin/gh/anshul-si/25/base 2025-09-07T07:35:59.9943765Z * [new branch] gh/anshul-si/25/head -> origin/gh/anshul-si/25/head 2025-09-07T07:35:59.9944288Z * [new branch] gh/anshul-si/25/orig -> origin/gh/anshul-si/25/orig 2025-09-07T07:35:59.9945134Z * [new branch] gh/anshul-si/26/base -> origin/gh/anshul-si/26/base 2025-09-07T07:35:59.9945573Z * [new branch] gh/anshul-si/26/head -> origin/gh/anshul-si/26/head 2025-09-07T07:35:59.9946107Z * [new branch] gh/anshul-si/26/orig -> origin/gh/anshul-si/26/orig 2025-09-07T07:35:59.9947013Z * [new branch] gh/anshul-si/27/base -> origin/gh/anshul-si/27/base 2025-09-07T07:35:59.9947435Z * [new branch] gh/anshul-si/27/head -> origin/gh/anshul-si/27/head 2025-09-07T07:35:59.9948151Z * [new branch] gh/anshul-si/27/orig -> origin/gh/anshul-si/27/orig 2025-09-07T07:35:59.9948956Z * [new branch] gh/anshul-si/28/base -> origin/gh/anshul-si/28/base 2025-09-07T07:35:59.9949509Z * [new branch] gh/anshul-si/28/head -> origin/gh/anshul-si/28/head 2025-09-07T07:35:59.9950007Z * [new branch] gh/anshul-si/28/orig -> origin/gh/anshul-si/28/orig 2025-09-07T07:35:59.9950721Z * [new branch] gh/anshul-si/29/base -> origin/gh/anshul-si/29/base 2025-09-07T07:35:59.9951362Z * [new branch] gh/anshul-si/29/head -> origin/gh/anshul-si/29/head 2025-09-07T07:35:59.9951797Z * [new branch] gh/anshul-si/29/orig -> origin/gh/anshul-si/29/orig 2025-09-07T07:35:59.9952592Z * [new branch] gh/anshul-si/3/base -> origin/gh/anshul-si/3/base 2025-09-07T07:35:59.9952952Z * [new branch] gh/anshul-si/3/head -> origin/gh/anshul-si/3/head 2025-09-07T07:35:59.9953701Z * [new branch] gh/anshul-si/4/base -> origin/gh/anshul-si/4/base 2025-09-07T07:35:59.9954124Z * [new branch] gh/anshul-si/4/head -> origin/gh/anshul-si/4/head 2025-09-07T07:35:59.9955143Z * [new branch] gh/anshul-si/5/base -> origin/gh/anshul-si/5/base 2025-09-07T07:35:59.9955500Z * [new branch] gh/anshul-si/5/head -> origin/gh/anshul-si/5/head 2025-09-07T07:35:59.9956758Z * [new branch] gh/aorenste/132/base -> origin/gh/aorenste/132/base 2025-09-07T07:35:59.9957454Z * [new branch] gh/aorenste/132/head -> origin/gh/aorenste/132/head 2025-09-07T07:35:59.9958548Z * [new branch] gh/bdhirsh/650/base -> origin/gh/bdhirsh/650/base 2025-09-07T07:35:59.9959403Z * [new branch] gh/bdhirsh/650/head -> origin/gh/bdhirsh/650/head 2025-09-07T07:35:59.9959668Z * [new branch] gh/bdhirsh/650/orig -> origin/gh/bdhirsh/650/orig 2025-09-07T07:35:59.9960558Z * [new branch] gh/bdhirsh/663/base -> origin/gh/bdhirsh/663/base 2025-09-07T07:35:59.9961256Z * [new branch] gh/bdhirsh/663/head -> origin/gh/bdhirsh/663/head 2025-09-07T07:35:59.9961614Z * [new branch] gh/bdhirsh/663/orig -> origin/gh/bdhirsh/663/orig 2025-09-07T07:35:59.9962570Z * [new branch] gh/bdhirsh/665/base -> origin/gh/bdhirsh/665/base 2025-09-07T07:35:59.9963114Z * [new branch] gh/bdhirsh/665/head -> origin/gh/bdhirsh/665/head 2025-09-07T07:35:59.9963770Z * [new branch] gh/bdhirsh/665/orig -> origin/gh/bdhirsh/665/orig 2025-09-07T07:35:59.9964800Z * [new branch] gh/bdhirsh/666/base -> origin/gh/bdhirsh/666/base 2025-09-07T07:35:59.9965349Z * [new branch] gh/bdhirsh/666/head -> origin/gh/bdhirsh/666/head 2025-09-07T07:35:59.9965906Z * [new branch] gh/bdhirsh/666/orig -> origin/gh/bdhirsh/666/orig 2025-09-07T07:35:59.9966883Z * [new branch] gh/bdhirsh/667/base -> origin/gh/bdhirsh/667/base 2025-09-07T07:35:59.9967307Z * [new branch] gh/bdhirsh/667/head -> origin/gh/bdhirsh/667/head 2025-09-07T07:35:59.9968090Z * [new branch] gh/bdhirsh/667/orig -> origin/gh/bdhirsh/667/orig 2025-09-07T07:35:59.9968910Z * [new branch] gh/bdhirsh/668/base -> origin/gh/bdhirsh/668/base 2025-09-07T07:35:59.9969336Z * [new branch] gh/bdhirsh/668/head -> origin/gh/bdhirsh/668/head 2025-09-07T07:35:59.9969985Z * [new branch] gh/bdhirsh/668/orig -> origin/gh/bdhirsh/668/orig 2025-09-07T07:35:59.9970837Z * [new branch] gh/bdhirsh/669/base -> origin/gh/bdhirsh/669/base 2025-09-07T07:35:59.9971260Z * [new branch] gh/bdhirsh/669/head -> origin/gh/bdhirsh/669/head 2025-09-07T07:35:59.9971912Z * [new branch] gh/bdhirsh/669/orig -> origin/gh/bdhirsh/669/orig 2025-09-07T07:35:59.9972830Z * [new branch] gh/bdhirsh/670/base -> origin/gh/bdhirsh/670/base 2025-09-07T07:35:59.9973491Z * [new branch] gh/bdhirsh/670/head -> origin/gh/bdhirsh/670/head 2025-09-07T07:35:59.9973932Z * [new branch] gh/bdhirsh/670/orig -> origin/gh/bdhirsh/670/orig 2025-09-07T07:35:59.9975064Z * [new branch] gh/benjaminglass1/100/base -> origin/gh/benjaminglass1/100/base 2025-09-07T07:35:59.9975472Z * [new branch] gh/benjaminglass1/100/head -> origin/gh/benjaminglass1/100/head 2025-09-07T07:35:59.9976213Z * [new branch] gh/benjaminglass1/100/orig -> origin/gh/benjaminglass1/100/orig 2025-09-07T07:35:59.9976979Z * [new branch] gh/benjaminglass1/101/base -> origin/gh/benjaminglass1/101/base 2025-09-07T07:35:59.9977380Z * [new branch] gh/benjaminglass1/101/head -> origin/gh/benjaminglass1/101/head 2025-09-07T07:35:59.9978178Z * [new branch] gh/benjaminglass1/101/orig -> origin/gh/benjaminglass1/101/orig 2025-09-07T07:35:59.9979005Z * [new branch] gh/benjaminglass1/102/base -> origin/gh/benjaminglass1/102/base 2025-09-07T07:35:59.9979571Z * [new branch] gh/benjaminglass1/102/head -> origin/gh/benjaminglass1/102/head 2025-09-07T07:35:59.9980129Z * [new branch] gh/benjaminglass1/102/orig -> origin/gh/benjaminglass1/102/orig 2025-09-07T07:35:59.9980941Z * [new branch] gh/benjaminglass1/103/base -> origin/gh/benjaminglass1/103/base 2025-09-07T07:35:59.9981331Z * [new branch] gh/benjaminglass1/103/head -> origin/gh/benjaminglass1/103/head 2025-09-07T07:35:59.9982036Z * [new branch] gh/benjaminglass1/103/orig -> origin/gh/benjaminglass1/103/orig 2025-09-07T07:35:59.9982753Z * [new branch] gh/benjaminglass1/104/base -> origin/gh/benjaminglass1/104/base 2025-09-07T07:35:59.9983143Z * [new branch] gh/benjaminglass1/104/head -> origin/gh/benjaminglass1/104/head 2025-09-07T07:35:59.9983800Z * [new branch] gh/benjaminglass1/104/orig -> origin/gh/benjaminglass1/104/orig 2025-09-07T07:35:59.9984555Z * [new branch] gh/benjaminglass1/105/base -> origin/gh/benjaminglass1/105/base 2025-09-07T07:35:59.9984964Z * [new branch] gh/benjaminglass1/105/head -> origin/gh/benjaminglass1/105/head 2025-09-07T07:35:59.9985630Z * [new branch] gh/benjaminglass1/105/orig -> origin/gh/benjaminglass1/105/orig 2025-09-07T07:35:59.9986402Z * [new branch] gh/benjaminglass1/106/base -> origin/gh/benjaminglass1/106/base 2025-09-07T07:35:59.9986961Z * [new branch] gh/benjaminglass1/106/head -> origin/gh/benjaminglass1/106/head 2025-09-07T07:35:59.9987667Z * [new branch] gh/benjaminglass1/106/orig -> origin/gh/benjaminglass1/106/orig 2025-09-07T07:35:59.9988390Z * [new branch] gh/benjaminglass1/79/base -> origin/gh/benjaminglass1/79/base 2025-09-07T07:35:59.9988811Z * [new branch] gh/benjaminglass1/79/head -> origin/gh/benjaminglass1/79/head 2025-09-07T07:35:59.9989382Z * [new branch] gh/benjaminglass1/79/orig -> origin/gh/benjaminglass1/79/orig 2025-09-07T07:35:59.9990205Z * [new branch] gh/benjaminglass1/86/base -> origin/gh/benjaminglass1/86/base 2025-09-07T07:35:59.9990651Z * [new branch] gh/benjaminglass1/86/head -> origin/gh/benjaminglass1/86/head 2025-09-07T07:35:59.9991304Z * [new branch] gh/benjaminglass1/86/orig -> origin/gh/benjaminglass1/86/orig 2025-09-07T07:35:59.9992013Z * [new branch] gh/benjaminglass1/89/base -> origin/gh/benjaminglass1/89/base 2025-09-07T07:35:59.9992452Z * [new branch] gh/benjaminglass1/89/head -> origin/gh/benjaminglass1/89/head 2025-09-07T07:35:59.9993478Z * [new branch] gh/benjaminglass1/89/orig -> origin/gh/benjaminglass1/89/orig 2025-09-07T07:35:59.9994176Z * [new branch] gh/benjaminglass1/91/base -> origin/gh/benjaminglass1/91/base 2025-09-07T07:35:59.9994823Z * [new branch] gh/benjaminglass1/91/head -> origin/gh/benjaminglass1/91/head 2025-09-07T07:35:59.9995263Z * [new branch] gh/benjaminglass1/91/orig -> origin/gh/benjaminglass1/91/orig 2025-09-07T07:35:59.9996106Z * [new branch] gh/benjaminglass1/93/base -> origin/gh/benjaminglass1/93/base 2025-09-07T07:35:59.9996547Z * [new branch] gh/benjaminglass1/93/head -> origin/gh/benjaminglass1/93/head 2025-09-07T07:35:59.9997283Z * [new branch] gh/benjaminglass1/93/orig -> origin/gh/benjaminglass1/93/orig 2025-09-07T07:35:59.9998044Z * [new branch] gh/benjaminglass1/95/base -> origin/gh/benjaminglass1/95/base 2025-09-07T07:35:59.9998475Z * [new branch] gh/benjaminglass1/95/head -> origin/gh/benjaminglass1/95/head 2025-09-07T07:35:59.9999086Z * [new branch] gh/benjaminglass1/95/orig -> origin/gh/benjaminglass1/95/orig 2025-09-07T07:35:59.9999895Z * [new branch] gh/benjaminglass1/97/base -> origin/gh/benjaminglass1/97/base 2025-09-07T07:36:00.0000302Z * [new branch] gh/benjaminglass1/97/head -> origin/gh/benjaminglass1/97/head 2025-09-07T07:36:00.0000980Z * [new branch] gh/benjaminglass1/97/orig -> origin/gh/benjaminglass1/97/orig 2025-09-07T07:36:00.0001725Z * [new branch] gh/benjaminglass1/99/base -> origin/gh/benjaminglass1/99/base 2025-09-07T07:36:00.0002283Z * [new branch] gh/benjaminglass1/99/head -> origin/gh/benjaminglass1/99/head 2025-09-07T07:36:00.0002954Z * [new branch] gh/benjaminglass1/99/orig -> origin/gh/benjaminglass1/99/orig 2025-09-07T07:36:00.0003936Z * [new branch] gh/bobrenjc93/514/base -> origin/gh/bobrenjc93/514/base 2025-09-07T07:36:00.0004318Z * [new branch] gh/bobrenjc93/514/head -> origin/gh/bobrenjc93/514/head 2025-09-07T07:36:00.0004984Z * [new branch] gh/bobrenjc93/514/orig -> origin/gh/bobrenjc93/514/orig 2025-09-07T07:36:00.0005698Z * [new branch] gh/bobrenjc93/521/base -> origin/gh/bobrenjc93/521/base 2025-09-07T07:36:00.0006348Z * [new branch] gh/bobrenjc93/521/head -> origin/gh/bobrenjc93/521/head 2025-09-07T07:36:00.0006791Z * [new branch] gh/bobrenjc93/521/orig -> origin/gh/bobrenjc93/521/orig 2025-09-07T07:36:00.0007600Z * [new branch] gh/bobrenjc93/522/base -> origin/gh/bobrenjc93/522/base 2025-09-07T07:36:00.0008031Z * [new branch] gh/bobrenjc93/522/head -> origin/gh/bobrenjc93/522/head 2025-09-07T07:36:00.0008677Z * [new branch] gh/bobrenjc93/522/orig -> origin/gh/bobrenjc93/522/orig 2025-09-07T07:36:00.0009477Z * [new branch] gh/bobrenjc93/525/base -> origin/gh/bobrenjc93/525/base 2025-09-07T07:36:00.0011400Z * [new branch] gh/bobrenjc93/525/head -> origin/gh/bobrenjc93/525/head 2025-09-07T07:36:00.0012203Z * [new branch] gh/bobrenjc93/525/orig -> origin/gh/bobrenjc93/525/orig 2025-09-07T07:36:00.0012346Z * [new branch] gh/bobrenjc93/526/base -> origin/gh/bobrenjc93/526/base 2025-09-07T07:36:00.0012469Z * [new branch] gh/bobrenjc93/526/head -> origin/gh/bobrenjc93/526/head 2025-09-07T07:36:00.0012605Z * [new branch] gh/bobrenjc93/526/orig -> origin/gh/bobrenjc93/526/orig 2025-09-07T07:36:00.0012843Z * [new branch] gh/bobrenjc93/527/base -> origin/gh/bobrenjc93/527/base 2025-09-07T07:36:00.0013436Z * [new branch] gh/bobrenjc93/527/head -> origin/gh/bobrenjc93/527/head 2025-09-07T07:36:00.0014029Z * [new branch] gh/bobrenjc93/527/orig -> origin/gh/bobrenjc93/527/orig 2025-09-07T07:36:00.0014732Z * [new branch] gh/bobrenjc93/528/base -> origin/gh/bobrenjc93/528/base 2025-09-07T07:36:00.0015435Z * [new branch] gh/bobrenjc93/528/head -> origin/gh/bobrenjc93/528/head 2025-09-07T07:36:00.0015866Z * [new branch] gh/bobrenjc93/528/orig -> origin/gh/bobrenjc93/528/orig 2025-09-07T07:36:00.0016775Z * [new branch] gh/bobrenjc93/529/base -> origin/gh/bobrenjc93/529/base 2025-09-07T07:36:00.0017204Z * [new branch] gh/bobrenjc93/529/head -> origin/gh/bobrenjc93/529/head 2025-09-07T07:36:00.0017865Z * [new branch] gh/bobrenjc93/529/orig -> origin/gh/bobrenjc93/529/orig 2025-09-07T07:36:00.0018627Z * [new branch] gh/bobrenjc93/535/base -> origin/gh/bobrenjc93/535/base 2025-09-07T07:36:00.0019029Z * [new branch] gh/bobrenjc93/535/head -> origin/gh/bobrenjc93/535/head 2025-09-07T07:36:00.0019658Z * [new branch] gh/bobrenjc93/535/orig -> origin/gh/bobrenjc93/535/orig 2025-09-07T07:36:00.0020425Z * [new branch] gh/bobrenjc93/537/base -> origin/gh/bobrenjc93/537/base 2025-09-07T07:36:00.0020978Z * [new branch] gh/bobrenjc93/537/head -> origin/gh/bobrenjc93/537/head 2025-09-07T07:36:00.0021411Z * [new branch] gh/bobrenjc93/537/orig -> origin/gh/bobrenjc93/537/orig 2025-09-07T07:36:00.0022461Z * [new branch] gh/bobrenjc93/539/base -> origin/gh/bobrenjc93/539/base 2025-09-07T07:36:00.0022908Z * [new branch] gh/bobrenjc93/539/head -> origin/gh/bobrenjc93/539/head 2025-09-07T07:36:00.0023584Z * [new branch] gh/bobrenjc93/539/orig -> origin/gh/bobrenjc93/539/orig 2025-09-07T07:36:00.0024425Z * [new branch] gh/bobrenjc93/540/base -> origin/gh/bobrenjc93/540/base 2025-09-07T07:36:00.0025054Z * [new branch] gh/bobrenjc93/540/head -> origin/gh/bobrenjc93/540/head 2025-09-07T07:36:00.0025633Z * [new branch] gh/bobrenjc93/540/orig -> origin/gh/bobrenjc93/540/orig 2025-09-07T07:36:00.0026401Z * [new branch] gh/bobrenjc93/541/base -> origin/gh/bobrenjc93/541/base 2025-09-07T07:36:00.0026819Z * [new branch] gh/bobrenjc93/541/head -> origin/gh/bobrenjc93/541/head 2025-09-07T07:36:00.0027448Z * [new branch] gh/bobrenjc93/541/orig -> origin/gh/bobrenjc93/541/orig 2025-09-07T07:36:00.0028131Z * [new branch] gh/bobrenjc93/542/base -> origin/gh/bobrenjc93/542/base 2025-09-07T07:36:00.0028575Z * [new branch] gh/bobrenjc93/542/head -> origin/gh/bobrenjc93/542/head 2025-09-07T07:36:00.0029245Z * [new branch] gh/bobrenjc93/542/orig -> origin/gh/bobrenjc93/542/orig 2025-09-07T07:36:00.0029980Z * [new branch] gh/bobrenjc93/543/base -> origin/gh/bobrenjc93/543/base 2025-09-07T07:36:00.0030378Z * [new branch] gh/bobrenjc93/543/head -> origin/gh/bobrenjc93/543/head 2025-09-07T07:36:00.0031037Z * [new branch] gh/bobrenjc93/543/orig -> origin/gh/bobrenjc93/543/orig 2025-09-07T07:36:00.0031768Z * [new branch] gh/bobrenjc93/544/base -> origin/gh/bobrenjc93/544/base 2025-09-07T07:36:00.0032597Z * [new branch] gh/bobrenjc93/544/head -> origin/gh/bobrenjc93/544/head 2025-09-07T07:36:00.0032737Z * [new branch] gh/bobrenjc93/544/orig -> origin/gh/bobrenjc93/544/orig 2025-09-07T07:36:00.0033782Z * [new branch] gh/bobrenjc93/545/base -> origin/gh/bobrenjc93/545/base 2025-09-07T07:36:00.0034580Z * [new branch] gh/bobrenjc93/545/head -> origin/gh/bobrenjc93/545/head 2025-09-07T07:36:00.0035184Z * [new branch] gh/bobrenjc93/545/orig -> origin/gh/bobrenjc93/545/orig 2025-09-07T07:36:00.0035918Z * [new branch] gh/bobrenjc93/546/base -> origin/gh/bobrenjc93/546/base 2025-09-07T07:36:00.0036489Z * [new branch] gh/bobrenjc93/546/head -> origin/gh/bobrenjc93/546/head 2025-09-07T07:36:00.0036934Z * [new branch] gh/bobrenjc93/546/orig -> origin/gh/bobrenjc93/546/orig 2025-09-07T07:36:00.0038212Z * [new branch] gh/bobrenjc93/547/base -> origin/gh/bobrenjc93/547/base 2025-09-07T07:36:00.0038768Z * [new branch] gh/bobrenjc93/547/head -> origin/gh/bobrenjc93/547/head 2025-09-07T07:36:00.0039356Z * [new branch] gh/bobrenjc93/547/orig -> origin/gh/bobrenjc93/547/orig 2025-09-07T07:36:00.0040086Z * [new branch] gh/bobrenjc93/548/base -> origin/gh/bobrenjc93/548/base 2025-09-07T07:36:00.0040503Z * [new branch] gh/bobrenjc93/548/head -> origin/gh/bobrenjc93/548/head 2025-09-07T07:36:00.0041085Z * [new branch] gh/bobrenjc93/548/orig -> origin/gh/bobrenjc93/548/orig 2025-09-07T07:36:00.0041733Z * [new branch] gh/bobrenjc93/549/base -> origin/gh/bobrenjc93/549/base 2025-09-07T07:36:00.0042338Z * [new branch] gh/bobrenjc93/549/head -> origin/gh/bobrenjc93/549/head 2025-09-07T07:36:00.0042887Z * [new branch] gh/bobrenjc93/549/orig -> origin/gh/bobrenjc93/549/orig 2025-09-07T07:36:00.0043943Z * [new branch] gh/bobrenjc93/550/base -> origin/gh/bobrenjc93/550/base 2025-09-07T07:36:00.0044587Z * [new branch] gh/bobrenjc93/550/head -> origin/gh/bobrenjc93/550/head 2025-09-07T07:36:00.0045171Z * [new branch] gh/bobrenjc93/550/orig -> origin/gh/bobrenjc93/550/orig 2025-09-07T07:36:00.0046091Z * [new branch] gh/bobrenjc93/551/base -> origin/gh/bobrenjc93/551/base 2025-09-07T07:36:00.0046643Z * [new branch] gh/bobrenjc93/551/head -> origin/gh/bobrenjc93/551/head 2025-09-07T07:36:00.0047082Z * [new branch] gh/bobrenjc93/551/orig -> origin/gh/bobrenjc93/551/orig 2025-09-07T07:36:00.0048080Z * [new branch] gh/bobrenjc93/552/base -> origin/gh/bobrenjc93/552/base 2025-09-07T07:36:00.0048654Z * [new branch] gh/bobrenjc93/552/head -> origin/gh/bobrenjc93/552/head 2025-09-07T07:36:00.0049109Z * [new branch] gh/bobrenjc93/552/orig -> origin/gh/bobrenjc93/552/orig 2025-09-07T07:36:00.0049850Z * [new branch] gh/bobrenjc93/553/base -> origin/gh/bobrenjc93/553/base 2025-09-07T07:36:00.0050349Z * [new branch] gh/bobrenjc93/553/head -> origin/gh/bobrenjc93/553/head 2025-09-07T07:36:00.0050900Z * [new branch] gh/bobrenjc93/553/orig -> origin/gh/bobrenjc93/553/orig 2025-09-07T07:36:00.0051566Z * [new branch] gh/bobrenjc93/554/base -> origin/gh/bobrenjc93/554/base 2025-09-07T07:36:00.0052019Z * [new branch] gh/bobrenjc93/554/head -> origin/gh/bobrenjc93/554/head 2025-09-07T07:36:00.0052574Z * [new branch] gh/bobrenjc93/554/orig -> origin/gh/bobrenjc93/554/orig 2025-09-07T07:36:00.0053501Z * [new branch] gh/bobrenjc93/555/base -> origin/gh/bobrenjc93/555/base 2025-09-07T07:36:00.0053903Z * [new branch] gh/bobrenjc93/555/head -> origin/gh/bobrenjc93/555/head 2025-09-07T07:36:00.0054527Z * [new branch] gh/bobrenjc93/555/orig -> origin/gh/bobrenjc93/555/orig 2025-09-07T07:36:00.0055344Z * [new branch] gh/bobrenjc93/556/base -> origin/gh/bobrenjc93/556/base 2025-09-07T07:36:00.0055835Z * [new branch] gh/bobrenjc93/556/head -> origin/gh/bobrenjc93/556/head 2025-09-07T07:36:00.0056375Z * [new branch] gh/bobrenjc93/556/orig -> origin/gh/bobrenjc93/556/orig 2025-09-07T07:36:00.0057350Z * [new branch] gh/briancoutinho/2/base -> origin/gh/briancoutinho/2/base 2025-09-07T07:36:00.0057849Z * [new branch] gh/briancoutinho/2/head -> origin/gh/briancoutinho/2/head 2025-09-07T07:36:00.0058981Z * [new branch] gh/c00w/23/base -> origin/gh/c00w/23/base 2025-09-07T07:36:00.0059525Z * [new branch] gh/c00w/23/head -> origin/gh/c00w/23/head 2025-09-07T07:36:00.0060389Z * [new branch] gh/c00w/48/base -> origin/gh/c00w/48/base 2025-09-07T07:36:00.0060813Z * [new branch] gh/c00w/48/head -> origin/gh/c00w/48/head 2025-09-07T07:36:00.0061421Z * [new branch] gh/c00w/48/orig -> origin/gh/c00w/48/orig 2025-09-07T07:36:00.0062331Z * [new branch] gh/c00w/53/base -> origin/gh/c00w/53/base 2025-09-07T07:36:00.0062743Z * [new branch] gh/c00w/53/head -> origin/gh/c00w/53/head 2025-09-07T07:36:00.0063445Z * [new branch] gh/c00w/53/orig -> origin/gh/c00w/53/orig 2025-09-07T07:36:00.0064095Z * [new branch] gh/c00w/54/base -> origin/gh/c00w/54/base 2025-09-07T07:36:00.0064628Z * [new branch] gh/c00w/54/head -> origin/gh/c00w/54/head 2025-09-07T07:36:00.0065165Z * [new branch] gh/c00w/54/orig -> origin/gh/c00w/54/orig 2025-09-07T07:36:00.0065953Z * [new branch] gh/c00w/55/base -> origin/gh/c00w/55/base 2025-09-07T07:36:00.0066598Z * [new branch] gh/c00w/55/head -> origin/gh/c00w/55/head 2025-09-07T07:36:00.0067129Z * [new branch] gh/c00w/55/orig -> origin/gh/c00w/55/orig 2025-09-07T07:36:00.0067839Z * [new branch] gh/c00w/56/base -> origin/gh/c00w/56/base 2025-09-07T07:36:00.0068451Z * [new branch] gh/c00w/56/head -> origin/gh/c00w/56/head 2025-09-07T07:36:00.0069043Z * [new branch] gh/c00w/56/orig -> origin/gh/c00w/56/orig 2025-09-07T07:36:00.0070048Z * [new branch] gh/clee2000/1/base -> origin/gh/clee2000/1/base 2025-09-07T07:36:00.0070724Z * [new branch] gh/clee2000/1/head -> origin/gh/clee2000/1/head 2025-09-07T07:36:00.0071130Z * [new branch] gh/clee2000/1/orig -> origin/gh/clee2000/1/orig 2025-09-07T07:36:00.0072147Z * [new branch] gh/coconutruben/1/base -> origin/gh/coconutruben/1/base 2025-09-07T07:36:00.0072778Z * [new branch] gh/coconutruben/1/head -> origin/gh/coconutruben/1/head 2025-09-07T07:36:00.0073836Z * [new branch] gh/coconutruben/11/base -> origin/gh/coconutruben/11/base 2025-09-07T07:36:00.0074416Z * [new branch] gh/coconutruben/11/head -> origin/gh/coconutruben/11/head 2025-09-07T07:36:00.0075093Z * [new branch] gh/coconutruben/11/orig -> origin/gh/coconutruben/11/orig 2025-09-07T07:36:00.0076285Z * [new branch] gh/coconutruben/12/base -> origin/gh/coconutruben/12/base 2025-09-07T07:36:00.0077050Z * [new branch] gh/coconutruben/12/head -> origin/gh/coconutruben/12/head 2025-09-07T07:36:00.0077800Z * [new branch] gh/coconutruben/12/orig -> origin/gh/coconutruben/12/orig 2025-09-07T07:36:00.0078663Z * [new branch] gh/coconutruben/13/base -> origin/gh/coconutruben/13/base 2025-09-07T07:36:00.0079269Z * [new branch] gh/coconutruben/13/head -> origin/gh/coconutruben/13/head 2025-09-07T07:36:00.0079824Z * [new branch] gh/coconutruben/13/orig -> origin/gh/coconutruben/13/orig 2025-09-07T07:36:00.0080692Z * [new branch] gh/coconutruben/14/base -> origin/gh/coconutruben/14/base 2025-09-07T07:36:00.0081345Z * [new branch] gh/coconutruben/14/head -> origin/gh/coconutruben/14/head 2025-09-07T07:36:00.0081891Z * [new branch] gh/coconutruben/14/orig -> origin/gh/coconutruben/14/orig 2025-09-07T07:36:00.0082818Z * [new branch] gh/coconutruben/15/base -> origin/gh/coconutruben/15/base 2025-09-07T07:36:00.0083432Z * [new branch] gh/coconutruben/15/head -> origin/gh/coconutruben/15/head 2025-09-07T07:36:00.0084035Z * [new branch] gh/coconutruben/15/orig -> origin/gh/coconutruben/15/orig 2025-09-07T07:36:00.0084891Z * [new branch] gh/coconutruben/16/base -> origin/gh/coconutruben/16/base 2025-09-07T07:36:00.0085449Z * [new branch] gh/coconutruben/16/head -> origin/gh/coconutruben/16/head 2025-09-07T07:36:00.0085891Z * [new branch] gh/coconutruben/16/orig -> origin/gh/coconutruben/16/orig 2025-09-07T07:36:00.0087155Z * [new branch] gh/coconutruben/17/base -> origin/gh/coconutruben/17/base 2025-09-07T07:36:00.0087775Z * [new branch] gh/coconutruben/17/head -> origin/gh/coconutruben/17/head 2025-09-07T07:36:00.0088212Z * [new branch] gh/coconutruben/17/orig -> origin/gh/coconutruben/17/orig 2025-09-07T07:36:00.0089086Z * [new branch] gh/coconutruben/18/base -> origin/gh/coconutruben/18/base 2025-09-07T07:36:00.0089786Z * [new branch] gh/coconutruben/18/head -> origin/gh/coconutruben/18/head 2025-09-07T07:36:00.0090168Z * [new branch] gh/coconutruben/18/orig -> origin/gh/coconutruben/18/orig 2025-09-07T07:36:00.0091202Z * [new branch] gh/coconutruben/19/base -> origin/gh/coconutruben/19/base 2025-09-07T07:36:00.0091798Z * [new branch] gh/coconutruben/19/head -> origin/gh/coconutruben/19/head 2025-09-07T07:36:00.0092363Z * [new branch] gh/coconutruben/19/orig -> origin/gh/coconutruben/19/orig 2025-09-07T07:36:00.0093246Z * [new branch] gh/coconutruben/20/base -> origin/gh/coconutruben/20/base 2025-09-07T07:36:00.0093692Z * [new branch] gh/coconutruben/20/head -> origin/gh/coconutruben/20/head 2025-09-07T07:36:00.0094411Z * [new branch] gh/coconutruben/20/orig -> origin/gh/coconutruben/20/orig 2025-09-07T07:36:00.0095439Z * [new branch] gh/coconutruben/21/base -> origin/gh/coconutruben/21/base 2025-09-07T07:36:00.0095893Z * [new branch] gh/coconutruben/21/head -> origin/gh/coconutruben/21/head 2025-09-07T07:36:00.0096562Z * [new branch] gh/coconutruben/21/orig -> origin/gh/coconutruben/21/orig 2025-09-07T07:36:00.0097318Z * [new branch] gh/coconutruben/22/base -> origin/gh/coconutruben/22/base 2025-09-07T07:36:00.0097752Z * [new branch] gh/coconutruben/22/head -> origin/gh/coconutruben/22/head 2025-09-07T07:36:00.0098516Z * [new branch] gh/coconutruben/22/orig -> origin/gh/coconutruben/22/orig 2025-09-07T07:36:00.0099319Z * [new branch] gh/coconutruben/24/base -> origin/gh/coconutruben/24/base 2025-09-07T07:36:00.0099946Z * [new branch] gh/coconutruben/24/head -> origin/gh/coconutruben/24/head 2025-09-07T07:36:00.0100506Z * [new branch] gh/coconutruben/24/orig -> origin/gh/coconutruben/24/orig 2025-09-07T07:36:00.0101592Z * [new branch] gh/coconutruben/25/base -> origin/gh/coconutruben/25/base 2025-09-07T07:36:00.0102508Z * [new branch] gh/coconutruben/25/head -> origin/gh/coconutruben/25/head 2025-09-07T07:36:00.0103229Z * [new branch] gh/coconutruben/25/orig -> origin/gh/coconutruben/25/orig 2025-09-07T07:36:00.0104114Z * [new branch] gh/coconutruben/28/base -> origin/gh/coconutruben/28/base 2025-09-07T07:36:00.0104707Z * [new branch] gh/coconutruben/28/head -> origin/gh/coconutruben/28/head 2025-09-07T07:36:00.0105342Z * [new branch] gh/coconutruben/28/orig -> origin/gh/coconutruben/28/orig 2025-09-07T07:36:00.0106284Z * [new branch] gh/coconutruben/29/base -> origin/gh/coconutruben/29/base 2025-09-07T07:36:00.0106845Z * [new branch] gh/coconutruben/29/head -> origin/gh/coconutruben/29/head 2025-09-07T07:36:00.0107439Z * [new branch] gh/coconutruben/29/orig -> origin/gh/coconutruben/29/orig 2025-09-07T07:36:00.0108355Z * [new branch] gh/coconutruben/30/base -> origin/gh/coconutruben/30/base 2025-09-07T07:36:00.0108907Z * [new branch] gh/coconutruben/30/head -> origin/gh/coconutruben/30/head 2025-09-07T07:36:00.0109510Z * [new branch] gh/coconutruben/30/orig -> origin/gh/coconutruben/30/orig 2025-09-07T07:36:00.0110650Z * [new branch] gh/coconutruben/31/base -> origin/gh/coconutruben/31/base 2025-09-07T07:36:00.0111204Z * [new branch] gh/coconutruben/31/head -> origin/gh/coconutruben/31/head 2025-09-07T07:36:00.0111765Z * [new branch] gh/coconutruben/31/orig -> origin/gh/coconutruben/31/orig 2025-09-07T07:36:00.0112849Z * [new branch] gh/coconutruben/32/base -> origin/gh/coconutruben/32/base 2025-09-07T07:36:00.0113487Z * [new branch] gh/coconutruben/32/head -> origin/gh/coconutruben/32/head 2025-09-07T07:36:00.0114116Z * [new branch] gh/coconutruben/32/orig -> origin/gh/coconutruben/32/orig 2025-09-07T07:36:00.0115082Z * [new branch] gh/coconutruben/33/base -> origin/gh/coconutruben/33/base 2025-09-07T07:36:00.0115634Z * [new branch] gh/coconutruben/33/head -> origin/gh/coconutruben/33/head 2025-09-07T07:36:00.0116326Z * [new branch] gh/coconutruben/33/orig -> origin/gh/coconutruben/33/orig 2025-09-07T07:36:00.0117058Z * [new branch] gh/coconutruben/34/base -> origin/gh/coconutruben/34/base 2025-09-07T07:36:00.0117663Z * [new branch] gh/coconutruben/34/head -> origin/gh/coconutruben/34/head 2025-09-07T07:36:00.0118013Z * [new branch] gh/coconutruben/34/orig -> origin/gh/coconutruben/34/orig 2025-09-07T07:36:00.0118848Z * [new branch] gh/coconutruben/35/base -> origin/gh/coconutruben/35/base 2025-09-07T07:36:00.0119408Z * [new branch] gh/coconutruben/35/head -> origin/gh/coconutruben/35/head 2025-09-07T07:36:00.0120031Z * [new branch] gh/coconutruben/35/orig -> origin/gh/coconutruben/35/orig 2025-09-07T07:36:00.0121900Z * [new branch] gh/coconutruben/36/base -> origin/gh/coconutruben/36/base 2025-09-07T07:36:00.0122798Z * [new branch] gh/coconutruben/36/head -> origin/gh/coconutruben/36/head 2025-09-07T07:36:00.0123968Z * [new branch] gh/coconutruben/36/orig -> origin/gh/coconutruben/36/orig 2025-09-07T07:36:00.0124978Z * [new branch] gh/coconutruben/37/base -> origin/gh/coconutruben/37/base 2025-09-07T07:36:00.0125400Z * [new branch] gh/coconutruben/37/head -> origin/gh/coconutruben/37/head 2025-09-07T07:36:00.0126063Z * [new branch] gh/coconutruben/37/orig -> origin/gh/coconutruben/37/orig 2025-09-07T07:36:00.0126947Z * [new branch] gh/coconutruben/38/base -> origin/gh/coconutruben/38/base 2025-09-07T07:36:00.0127622Z * [new branch] gh/coconutruben/38/head -> origin/gh/coconutruben/38/head 2025-09-07T07:36:00.0128263Z * [new branch] gh/coconutruben/38/orig -> origin/gh/coconutruben/38/orig 2025-09-07T07:36:00.0129529Z * [new branch] gh/coconutruben/39/base -> origin/gh/coconutruben/39/base 2025-09-07T07:36:00.0129952Z * [new branch] gh/coconutruben/39/head -> origin/gh/coconutruben/39/head 2025-09-07T07:36:00.0130592Z * [new branch] gh/coconutruben/39/orig -> origin/gh/coconutruben/39/orig 2025-09-07T07:36:00.0131639Z * [new branch] gh/coconutruben/40/base -> origin/gh/coconutruben/40/base 2025-09-07T07:36:00.0132075Z * [new branch] gh/coconutruben/40/head -> origin/gh/coconutruben/40/head 2025-09-07T07:36:00.0132670Z * [new branch] gh/coconutruben/40/orig -> origin/gh/coconutruben/40/orig 2025-09-07T07:36:00.0133654Z * [new branch] gh/coconutruben/41/base -> origin/gh/coconutruben/41/base 2025-09-07T07:36:00.0134547Z * [new branch] gh/coconutruben/41/head -> origin/gh/coconutruben/41/head 2025-09-07T07:36:00.0135105Z * [new branch] gh/coconutruben/41/orig -> origin/gh/coconutruben/41/orig 2025-09-07T07:36:00.0136051Z * [new branch] gh/coconutruben/42/base -> origin/gh/coconutruben/42/base 2025-09-07T07:36:00.0136613Z * [new branch] gh/coconutruben/42/head -> origin/gh/coconutruben/42/head 2025-09-07T07:36:00.0137204Z * [new branch] gh/coconutruben/42/orig -> origin/gh/coconutruben/42/orig 2025-09-07T07:36:00.0138120Z * [new branch] gh/coconutruben/43/base -> origin/gh/coconutruben/43/base 2025-09-07T07:36:00.0138818Z * [new branch] gh/coconutruben/43/head -> origin/gh/coconutruben/43/head 2025-09-07T07:36:00.0139314Z * [new branch] gh/coconutruben/43/orig -> origin/gh/coconutruben/43/orig 2025-09-07T07:36:00.0140423Z * [new branch] gh/coconutruben/44/base -> origin/gh/coconutruben/44/base 2025-09-07T07:36:00.0141042Z * [new branch] gh/coconutruben/44/head -> origin/gh/coconutruben/44/head 2025-09-07T07:36:00.0141671Z * [new branch] gh/coconutruben/44/orig -> origin/gh/coconutruben/44/orig 2025-09-07T07:36:00.0142610Z * [new branch] gh/coconutruben/45/base -> origin/gh/coconutruben/45/base 2025-09-07T07:36:00.0143170Z * [new branch] gh/coconutruben/45/head -> origin/gh/coconutruben/45/head 2025-09-07T07:36:00.0143763Z * [new branch] gh/coconutruben/45/orig -> origin/gh/coconutruben/45/orig 2025-09-07T07:36:00.0144568Z * [new branch] gh/coconutruben/46/base -> origin/gh/coconutruben/46/base 2025-09-07T07:36:00.0145136Z * [new branch] gh/coconutruben/46/head -> origin/gh/coconutruben/46/head 2025-09-07T07:36:00.0145701Z * [new branch] gh/coconutruben/46/orig -> origin/gh/coconutruben/46/orig 2025-09-07T07:36:00.0146654Z * [new branch] gh/coconutruben/47/base -> origin/gh/coconutruben/47/base 2025-09-07T07:36:00.0147201Z * [new branch] gh/coconutruben/47/head -> origin/gh/coconutruben/47/head 2025-09-07T07:36:00.0147789Z * [new branch] gh/coconutruben/47/orig -> origin/gh/coconutruben/47/orig 2025-09-07T07:36:00.0148897Z * [new branch] gh/coconutruben/48/base -> origin/gh/coconutruben/48/base 2025-09-07T07:36:00.0149486Z * [new branch] gh/coconutruben/48/head -> origin/gh/coconutruben/48/head 2025-09-07T07:36:00.0150130Z * [new branch] gh/coconutruben/48/orig -> origin/gh/coconutruben/48/orig 2025-09-07T07:36:00.0151178Z * [new branch] gh/coconutruben/49/base -> origin/gh/coconutruben/49/base 2025-09-07T07:36:00.0151754Z * [new branch] gh/coconutruben/49/head -> origin/gh/coconutruben/49/head 2025-09-07T07:36:00.0152342Z * [new branch] gh/coconutruben/49/orig -> origin/gh/coconutruben/49/orig 2025-09-07T07:36:00.0153485Z * [new branch] gh/coconutruben/50/base -> origin/gh/coconutruben/50/base 2025-09-07T07:36:00.0154123Z * [new branch] gh/coconutruben/50/head -> origin/gh/coconutruben/50/head 2025-09-07T07:36:00.0154719Z * [new branch] gh/coconutruben/50/orig -> origin/gh/coconutruben/50/orig 2025-09-07T07:36:00.0155444Z * [new branch] gh/coconutruben/51/base -> origin/gh/coconutruben/51/base 2025-09-07T07:36:00.0156018Z * [new branch] gh/coconutruben/51/head -> origin/gh/coconutruben/51/head 2025-09-07T07:36:00.0156617Z * [new branch] gh/coconutruben/51/orig -> origin/gh/coconutruben/51/orig 2025-09-07T07:36:00.0157710Z * [new branch] gh/coconutruben/52/base -> origin/gh/coconutruben/52/base 2025-09-07T07:36:00.0158329Z * [new branch] gh/coconutruben/52/head -> origin/gh/coconutruben/52/head 2025-09-07T07:36:00.0158960Z * [new branch] gh/coconutruben/52/orig -> origin/gh/coconutruben/52/orig 2025-09-07T07:36:00.0159812Z * [new branch] gh/coconutruben/53/base -> origin/gh/coconutruben/53/base 2025-09-07T07:36:00.0160405Z * [new branch] gh/coconutruben/53/head -> origin/gh/coconutruben/53/head 2025-09-07T07:36:00.0160969Z * [new branch] gh/coconutruben/53/orig -> origin/gh/coconutruben/53/orig 2025-09-07T07:36:00.0161848Z * [new branch] gh/coconutruben/54/base -> origin/gh/coconutruben/54/base 2025-09-07T07:36:00.0162468Z * [new branch] gh/coconutruben/54/head -> origin/gh/coconutruben/54/head 2025-09-07T07:36:00.0163031Z * [new branch] gh/coconutruben/54/orig -> origin/gh/coconutruben/54/orig 2025-09-07T07:36:00.0163895Z * [new branch] gh/coconutruben/55/base -> origin/gh/coconutruben/55/base 2025-09-07T07:36:00.0164346Z * [new branch] gh/coconutruben/55/head -> origin/gh/coconutruben/55/head 2025-09-07T07:36:00.0165051Z * [new branch] gh/coconutruben/55/orig -> origin/gh/coconutruben/55/orig 2025-09-07T07:36:00.0165908Z * [new branch] gh/coconutruben/56/base -> origin/gh/coconutruben/56/base 2025-09-07T07:36:00.0166568Z * [new branch] gh/coconutruben/56/head -> origin/gh/coconutruben/56/head 2025-09-07T07:36:00.0167169Z * [new branch] gh/coconutruben/56/orig -> origin/gh/coconutruben/56/orig 2025-09-07T07:36:00.0168011Z * [new branch] gh/coconutruben/57/base -> origin/gh/coconutruben/57/base 2025-09-07T07:36:00.0168671Z * [new branch] gh/coconutruben/57/head -> origin/gh/coconutruben/57/head 2025-09-07T07:36:00.0169300Z * [new branch] gh/coconutruben/57/orig -> origin/gh/coconutruben/57/orig 2025-09-07T07:36:00.0170981Z * [new branch] gh/coconutruben/58/base -> origin/gh/coconutruben/58/base 2025-09-07T07:36:00.0176049Z * [new branch] gh/coconutruben/58/head -> origin/gh/coconutruben/58/head 2025-09-07T07:36:00.0176199Z * [new branch] gh/coconutruben/58/orig -> origin/gh/coconutruben/58/orig 2025-09-07T07:36:00.0176341Z * [new branch] gh/coconutruben/59/base -> origin/gh/coconutruben/59/base 2025-09-07T07:36:00.0176482Z * [new branch] gh/coconutruben/59/head -> origin/gh/coconutruben/59/head 2025-09-07T07:36:00.0176620Z * [new branch] gh/coconutruben/59/orig -> origin/gh/coconutruben/59/orig 2025-09-07T07:36:00.0176756Z * [new branch] gh/coconutruben/60/base -> origin/gh/coconutruben/60/base 2025-09-07T07:36:00.0176899Z * [new branch] gh/coconutruben/60/head -> origin/gh/coconutruben/60/head 2025-09-07T07:36:00.0177031Z * [new branch] gh/coconutruben/60/orig -> origin/gh/coconutruben/60/orig 2025-09-07T07:36:00.0177187Z * [new branch] gh/coconutruben/61/base -> origin/gh/coconutruben/61/base 2025-09-07T07:36:00.0177327Z * [new branch] gh/coconutruben/61/head -> origin/gh/coconutruben/61/head 2025-09-07T07:36:00.0177639Z * [new branch] gh/coconutruben/61/orig -> origin/gh/coconutruben/61/orig 2025-09-07T07:36:00.0178838Z * [new branch] gh/coconutruben/62/base -> origin/gh/coconutruben/62/base 2025-09-07T07:36:00.0179248Z * [new branch] gh/coconutruben/62/head -> origin/gh/coconutruben/62/head 2025-09-07T07:36:00.0179726Z * [new branch] gh/coconutruben/62/orig -> origin/gh/coconutruben/62/orig 2025-09-07T07:36:00.0180621Z * [new branch] gh/coconutruben/63/base -> origin/gh/coconutruben/63/base 2025-09-07T07:36:00.0181376Z * [new branch] gh/coconutruben/63/head -> origin/gh/coconutruben/63/head 2025-09-07T07:36:00.0181912Z * [new branch] gh/coconutruben/63/orig -> origin/gh/coconutruben/63/orig 2025-09-07T07:36:00.0182768Z * [new branch] gh/coconutruben/64/base -> origin/gh/coconutruben/64/base 2025-09-07T07:36:00.0183460Z * [new branch] gh/coconutruben/64/head -> origin/gh/coconutruben/64/head 2025-09-07T07:36:00.0183984Z * [new branch] gh/coconutruben/64/orig -> origin/gh/coconutruben/64/orig 2025-09-07T07:36:00.0184911Z * [new branch] gh/coconutruben/65/base -> origin/gh/coconutruben/65/base 2025-09-07T07:36:00.0185457Z * [new branch] gh/coconutruben/65/head -> origin/gh/coconutruben/65/head 2025-09-07T07:36:00.0185964Z * [new branch] gh/coconutruben/65/orig -> origin/gh/coconutruben/65/orig 2025-09-07T07:36:00.0186876Z * [new branch] gh/coconutruben/66/base -> origin/gh/coconutruben/66/base 2025-09-07T07:36:00.0187319Z * [new branch] gh/coconutruben/66/head -> origin/gh/coconutruben/66/head 2025-09-07T07:36:00.0187816Z * [new branch] gh/coconutruben/66/orig -> origin/gh/coconutruben/66/orig 2025-09-07T07:36:00.0189250Z * [new branch] gh/codingwithsurya/12/base -> origin/gh/codingwithsurya/12/base 2025-09-07T07:36:00.0189862Z * [new branch] gh/codingwithsurya/12/head -> origin/gh/codingwithsurya/12/head 2025-09-07T07:36:00.0190623Z * [new branch] gh/codingwithsurya/12/orig -> origin/gh/codingwithsurya/12/orig 2025-09-07T07:36:00.0191402Z * [new branch] gh/codingwithsurya/14/base -> origin/gh/codingwithsurya/14/base 2025-09-07T07:36:00.0192011Z * [new branch] gh/codingwithsurya/14/head -> origin/gh/codingwithsurya/14/head 2025-09-07T07:36:00.0192603Z * [new branch] gh/codingwithsurya/14/orig -> origin/gh/codingwithsurya/14/orig 2025-09-07T07:36:00.0193602Z * [new branch] gh/codingwithsurya/15/base -> origin/gh/codingwithsurya/15/base 2025-09-07T07:36:00.0194099Z * [new branch] gh/codingwithsurya/15/head -> origin/gh/codingwithsurya/15/head 2025-09-07T07:36:00.0194899Z * [new branch] gh/codingwithsurya/15/orig -> origin/gh/codingwithsurya/15/orig 2025-09-07T07:36:00.0195678Z * [new branch] gh/codingwithsurya/16/base -> origin/gh/codingwithsurya/16/base 2025-09-07T07:36:00.0196186Z * [new branch] gh/codingwithsurya/16/head -> origin/gh/codingwithsurya/16/head 2025-09-07T07:36:00.0196737Z * [new branch] gh/codingwithsurya/16/orig -> origin/gh/codingwithsurya/16/orig 2025-09-07T07:36:00.0197806Z * [new branch] gh/codingwithsurya/17/base -> origin/gh/codingwithsurya/17/base 2025-09-07T07:36:00.0198261Z * [new branch] gh/codingwithsurya/17/head -> origin/gh/codingwithsurya/17/head 2025-09-07T07:36:00.0198811Z * [new branch] gh/codingwithsurya/17/orig -> origin/gh/codingwithsurya/17/orig 2025-09-07T07:36:00.0199771Z * [new branch] gh/codingwithsurya/18/base -> origin/gh/codingwithsurya/18/base 2025-09-07T07:36:00.0200280Z * [new branch] gh/codingwithsurya/18/head -> origin/gh/codingwithsurya/18/head 2025-09-07T07:36:00.0200803Z * [new branch] gh/codingwithsurya/18/orig -> origin/gh/codingwithsurya/18/orig 2025-09-07T07:36:00.0201929Z * [new branch] gh/codingwithsurya/19/base -> origin/gh/codingwithsurya/19/base 2025-09-07T07:36:00.0202395Z * [new branch] gh/codingwithsurya/19/head -> origin/gh/codingwithsurya/19/head 2025-09-07T07:36:00.0202894Z * [new branch] gh/codingwithsurya/19/orig -> origin/gh/codingwithsurya/19/orig 2025-09-07T07:36:00.0203800Z * [new branch] gh/codingwithsurya/20/base -> origin/gh/codingwithsurya/20/base 2025-09-07T07:36:00.0204233Z * [new branch] gh/codingwithsurya/20/head -> origin/gh/codingwithsurya/20/head 2025-09-07T07:36:00.0204801Z * [new branch] gh/codingwithsurya/20/orig -> origin/gh/codingwithsurya/20/orig 2025-09-07T07:36:00.0205805Z * [new branch] gh/codingwithsurya/21/base -> origin/gh/codingwithsurya/21/base 2025-09-07T07:36:00.0206323Z * [new branch] gh/codingwithsurya/21/head -> origin/gh/codingwithsurya/21/head 2025-09-07T07:36:00.0206847Z * [new branch] gh/codingwithsurya/21/orig -> origin/gh/codingwithsurya/21/orig 2025-09-07T07:36:00.0207943Z * [new branch] gh/colinchan15/1/base -> origin/gh/colinchan15/1/base 2025-09-07T07:36:00.0208412Z * [new branch] gh/colinchan15/1/head -> origin/gh/colinchan15/1/head 2025-09-07T07:36:00.0209141Z * [new branch] gh/colinchan15/2/base -> origin/gh/colinchan15/2/base 2025-09-07T07:36:00.0209528Z * [new branch] gh/colinchan15/2/head -> origin/gh/colinchan15/2/head 2025-09-07T07:36:00.0210256Z * [new branch] gh/colinchan15/3/base -> origin/gh/colinchan15/3/base 2025-09-07T07:36:00.0210648Z * [new branch] gh/colinchan15/3/head -> origin/gh/colinchan15/3/head 2025-09-07T07:36:00.0211555Z * [new branch] gh/colinchan15/6/base -> origin/gh/colinchan15/6/base 2025-09-07T07:36:00.0212023Z * [new branch] gh/colinchan15/6/head -> origin/gh/colinchan15/6/head 2025-09-07T07:36:00.0213123Z * [new branch] gh/davidberard98/382/base -> origin/gh/davidberard98/382/base 2025-09-07T07:36:00.0213756Z * [new branch] gh/davidberard98/382/head -> origin/gh/davidberard98/382/head 2025-09-07T07:36:00.0214407Z * [new branch] gh/davidberard98/382/orig -> origin/gh/davidberard98/382/orig 2025-09-07T07:36:00.0215267Z * [new branch] gh/davidberard98/386/base -> origin/gh/davidberard98/386/base 2025-09-07T07:36:00.0215780Z * [new branch] gh/davidberard98/386/head -> origin/gh/davidberard98/386/head 2025-09-07T07:36:00.0216334Z * [new branch] gh/davidberard98/386/orig -> origin/gh/davidberard98/386/orig 2025-09-07T07:36:00.0217239Z * [new branch] gh/davidberard98/391/base -> origin/gh/davidberard98/391/base 2025-09-07T07:36:00.0217652Z * [new branch] gh/davidberard98/391/head -> origin/gh/davidberard98/391/head 2025-09-07T07:36:00.0218167Z * [new branch] gh/davidberard98/391/orig -> origin/gh/davidberard98/391/orig 2025-09-07T07:36:00.0219052Z * [new branch] gh/davidberard98/392/base -> origin/gh/davidberard98/392/base 2025-09-07T07:36:00.0219505Z * [new branch] gh/davidberard98/392/head -> origin/gh/davidberard98/392/head 2025-09-07T07:36:00.0220046Z * [new branch] gh/davidberard98/392/orig -> origin/gh/davidberard98/392/orig 2025-09-07T07:36:00.0221001Z * [new branch] gh/davidberard98/394/base -> origin/gh/davidberard98/394/base 2025-09-07T07:36:00.0221602Z * [new branch] gh/davidberard98/394/head -> origin/gh/davidberard98/394/head 2025-09-07T07:36:00.0222214Z * [new branch] gh/davidberard98/394/orig -> origin/gh/davidberard98/394/orig 2025-09-07T07:36:00.0223072Z * [new branch] gh/davidberard98/396/base -> origin/gh/davidberard98/396/base 2025-09-07T07:36:00.0223523Z * [new branch] gh/davidberard98/396/head -> origin/gh/davidberard98/396/head 2025-09-07T07:36:00.0224102Z * [new branch] gh/davidberard98/396/orig -> origin/gh/davidberard98/396/orig 2025-09-07T07:36:00.0225133Z * [new branch] gh/davidberard98/397/base -> origin/gh/davidberard98/397/base 2025-09-07T07:36:00.0225581Z * [new branch] gh/davidberard98/397/head -> origin/gh/davidberard98/397/head 2025-09-07T07:36:00.0226132Z * [new branch] gh/davidberard98/397/orig -> origin/gh/davidberard98/397/orig 2025-09-07T07:36:00.0226981Z * [new branch] gh/davidberard98/398/base -> origin/gh/davidberard98/398/base 2025-09-07T07:36:00.0227390Z * [new branch] gh/davidberard98/398/head -> origin/gh/davidberard98/398/head 2025-09-07T07:36:00.0227954Z * [new branch] gh/davidberard98/398/orig -> origin/gh/davidberard98/398/orig 2025-09-07T07:36:00.0228840Z * [new branch] gh/davidberard98/399/base -> origin/gh/davidberard98/399/base 2025-09-07T07:36:00.0229332Z * [new branch] gh/davidberard98/399/head -> origin/gh/davidberard98/399/head 2025-09-07T07:36:00.0229869Z * [new branch] gh/davidberard98/399/orig -> origin/gh/davidberard98/399/orig 2025-09-07T07:36:00.0230886Z * [new branch] gh/davidberard98/400/base -> origin/gh/davidberard98/400/base 2025-09-07T07:36:00.0231496Z * [new branch] gh/davidberard98/400/head -> origin/gh/davidberard98/400/head 2025-09-07T07:36:00.0232035Z * [new branch] gh/davidberard98/400/orig -> origin/gh/davidberard98/400/orig 2025-09-07T07:36:00.0232838Z * [new branch] gh/davidberard98/401/base -> origin/gh/davidberard98/401/base 2025-09-07T07:36:00.0233286Z * [new branch] gh/davidberard98/401/head -> origin/gh/davidberard98/401/head 2025-09-07T07:36:00.0233878Z * [new branch] gh/davidberard98/401/orig -> origin/gh/davidberard98/401/orig 2025-09-07T07:36:00.0234688Z * [new branch] gh/davidberard98/402/base -> origin/gh/davidberard98/402/base 2025-09-07T07:36:00.0235136Z * [new branch] gh/davidberard98/402/head -> origin/gh/davidberard98/402/head 2025-09-07T07:36:00.0235675Z * [new branch] gh/davidberard98/402/orig -> origin/gh/davidberard98/402/orig 2025-09-07T07:36:00.0236527Z * [new branch] gh/davidberard98/403/base -> origin/gh/davidberard98/403/base 2025-09-07T07:36:00.0236986Z * [new branch] gh/davidberard98/403/head -> origin/gh/davidberard98/403/head 2025-09-07T07:36:00.0237506Z * [new branch] gh/davidberard98/403/orig -> origin/gh/davidberard98/403/orig 2025-09-07T07:36:00.0238480Z * [new branch] gh/davidberard98/404/base -> origin/gh/davidberard98/404/base 2025-09-07T07:36:00.0238916Z * [new branch] gh/davidberard98/404/head -> origin/gh/davidberard98/404/head 2025-09-07T07:36:00.0239405Z * [new branch] gh/davidberard98/404/orig -> origin/gh/davidberard98/404/orig 2025-09-07T07:36:00.0240238Z * [new branch] gh/davidberard98/405/base -> origin/gh/davidberard98/405/base 2025-09-07T07:36:00.0240840Z * [new branch] gh/davidberard98/405/head -> origin/gh/davidberard98/405/head 2025-09-07T07:36:00.0241379Z * [new branch] gh/davidberard98/405/orig -> origin/gh/davidberard98/405/orig 2025-09-07T07:36:00.0242278Z * [new branch] gh/davidberard98/406/base -> origin/gh/davidberard98/406/base 2025-09-07T07:36:00.0242847Z * [new branch] gh/davidberard98/406/head -> origin/gh/davidberard98/406/head 2025-09-07T07:36:00.0243593Z * [new branch] gh/davidberard98/406/orig -> origin/gh/davidberard98/406/orig 2025-09-07T07:36:00.0244976Z * [new branch] gh/davidberard98/407/base -> origin/gh/davidberard98/407/base 2025-09-07T07:36:00.0245314Z * [new branch] gh/davidberard98/407/head -> origin/gh/davidberard98/407/head 2025-09-07T07:36:00.0245725Z * [new branch] gh/davidberard98/407/orig -> origin/gh/davidberard98/407/orig 2025-09-07T07:36:00.0246636Z * [new branch] gh/davidberard98/408/base -> origin/gh/davidberard98/408/base 2025-09-07T07:36:00.0247073Z * [new branch] gh/davidberard98/408/head -> origin/gh/davidberard98/408/head 2025-09-07T07:36:00.0247589Z * [new branch] gh/davidberard98/408/orig -> origin/gh/davidberard98/408/orig 2025-09-07T07:36:00.0248375Z * [new branch] gh/davidberard98/409/base -> origin/gh/davidberard98/409/base 2025-09-07T07:36:00.0248903Z * [new branch] gh/davidberard98/409/head -> origin/gh/davidberard98/409/head 2025-09-07T07:36:00.0249534Z * [new branch] gh/davidberard98/409/orig -> origin/gh/davidberard98/409/orig 2025-09-07T07:36:00.0250675Z * [new branch] gh/desertfire/594/base -> origin/gh/desertfire/594/base 2025-09-07T07:36:00.0251140Z * [new branch] gh/desertfire/594/head -> origin/gh/desertfire/594/head 2025-09-07T07:36:00.0251758Z * [new branch] gh/desertfire/594/orig -> origin/gh/desertfire/594/orig 2025-09-07T07:36:00.0252559Z * [new branch] gh/desertfire/595/base -> origin/gh/desertfire/595/base 2025-09-07T07:36:00.0252961Z * [new branch] gh/desertfire/595/head -> origin/gh/desertfire/595/head 2025-09-07T07:36:00.0253492Z * [new branch] gh/desertfire/595/orig -> origin/gh/desertfire/595/orig 2025-09-07T07:36:00.0254410Z * [new branch] gh/desertfire/597/base -> origin/gh/desertfire/597/base 2025-09-07T07:36:00.0254881Z * [new branch] gh/desertfire/597/head -> origin/gh/desertfire/597/head 2025-09-07T07:36:00.0255467Z * [new branch] gh/desertfire/597/orig -> origin/gh/desertfire/597/orig 2025-09-07T07:36:00.0256508Z * [new branch] gh/dharakk/1/base -> origin/gh/dharakk/1/base 2025-09-07T07:36:00.0256983Z * [new branch] gh/dharakk/1/head -> origin/gh/dharakk/1/head 2025-09-07T07:36:00.0258007Z * [new branch] gh/drisspg/149/base -> origin/gh/drisspg/149/base 2025-09-07T07:36:00.0258434Z * [new branch] gh/drisspg/149/head -> origin/gh/drisspg/149/head 2025-09-07T07:36:00.0258970Z * [new branch] gh/drisspg/149/orig -> origin/gh/drisspg/149/orig 2025-09-07T07:36:00.0259775Z * [new branch] gh/drisspg/159/base -> origin/gh/drisspg/159/base 2025-09-07T07:36:00.0260326Z * [new branch] gh/drisspg/159/head -> origin/gh/drisspg/159/head 2025-09-07T07:36:00.0260856Z * [new branch] gh/drisspg/159/orig -> origin/gh/drisspg/159/orig 2025-09-07T07:36:00.0262113Z * [new branch] gh/drisspg/166/base -> origin/gh/drisspg/166/base 2025-09-07T07:36:00.0262516Z * [new branch] gh/drisspg/166/head -> origin/gh/drisspg/166/head 2025-09-07T07:36:00.0263089Z * [new branch] gh/drisspg/166/orig -> origin/gh/drisspg/166/orig 2025-09-07T07:36:00.0263898Z * [new branch] gh/drisspg/170/base -> origin/gh/drisspg/170/base 2025-09-07T07:36:00.0264356Z * [new branch] gh/drisspg/170/head -> origin/gh/drisspg/170/head 2025-09-07T07:36:00.0264883Z * [new branch] gh/drisspg/170/orig -> origin/gh/drisspg/170/orig 2025-09-07T07:36:00.0265672Z * [new branch] gh/drisspg/173/base -> origin/gh/drisspg/173/base 2025-09-07T07:36:00.0266102Z * [new branch] gh/drisspg/173/head -> origin/gh/drisspg/173/head 2025-09-07T07:36:00.0266616Z * [new branch] gh/drisspg/173/orig -> origin/gh/drisspg/173/orig 2025-09-07T07:36:00.0267425Z * [new branch] gh/drisspg/177/base -> origin/gh/drisspg/177/base 2025-09-07T07:36:00.0267881Z * [new branch] gh/drisspg/177/head -> origin/gh/drisspg/177/head 2025-09-07T07:36:00.0268433Z * [new branch] gh/drisspg/177/orig -> origin/gh/drisspg/177/orig 2025-09-07T07:36:00.0273448Z * [new branch] gh/drisspg/178/base -> origin/gh/drisspg/178/base 2025-09-07T07:36:00.0274072Z * [new branch] gh/drisspg/178/head -> origin/gh/drisspg/178/head 2025-09-07T07:36:00.0274556Z * [new branch] gh/drisspg/178/orig -> origin/gh/drisspg/178/orig 2025-09-07T07:36:00.0275418Z * [new branch] gh/drisspg/180/base -> origin/gh/drisspg/180/base 2025-09-07T07:36:00.0275904Z * [new branch] gh/drisspg/180/head -> origin/gh/drisspg/180/head 2025-09-07T07:36:00.0276473Z * [new branch] gh/drisspg/180/orig -> origin/gh/drisspg/180/orig 2025-09-07T07:36:00.0277263Z * [new branch] gh/drisspg/181/base -> origin/gh/drisspg/181/base 2025-09-07T07:36:00.0277701Z * [new branch] gh/drisspg/181/head -> origin/gh/drisspg/181/head 2025-09-07T07:36:00.0278286Z * [new branch] gh/drisspg/181/orig -> origin/gh/drisspg/181/orig 2025-09-07T07:36:00.0279166Z * [new branch] gh/drisspg/182/base -> origin/gh/drisspg/182/base 2025-09-07T07:36:00.0279645Z * [new branch] gh/drisspg/182/head -> origin/gh/drisspg/182/head 2025-09-07T07:36:00.0280442Z * [new branch] gh/drisspg/183/base -> origin/gh/drisspg/183/base 2025-09-07T07:36:00.0280846Z * [new branch] gh/drisspg/183/head -> origin/gh/drisspg/183/head 2025-09-07T07:36:00.0281567Z * [new branch] gh/drisspg/184/base -> origin/gh/drisspg/184/base 2025-09-07T07:36:00.0281953Z * [new branch] gh/drisspg/184/head -> origin/gh/drisspg/184/head 2025-09-07T07:36:00.0282821Z * [new branch] gh/drisspg/185/base -> origin/gh/drisspg/185/base 2025-09-07T07:36:00.0283376Z * [new branch] gh/drisspg/185/head -> origin/gh/drisspg/185/head 2025-09-07T07:36:00.0284532Z * [new branch] gh/drisspg/186/base -> origin/gh/drisspg/186/base 2025-09-07T07:36:00.0284982Z * [new branch] gh/drisspg/186/head -> origin/gh/drisspg/186/head 2025-09-07T07:36:00.0285525Z * [new branch] gh/drisspg/186/orig -> origin/gh/drisspg/186/orig 2025-09-07T07:36:00.0286352Z * [new branch] gh/drisspg/187/base -> origin/gh/drisspg/187/base 2025-09-07T07:36:00.0286811Z * [new branch] gh/drisspg/187/head -> origin/gh/drisspg/187/head 2025-09-07T07:36:00.0287349Z * [new branch] gh/drisspg/187/orig -> origin/gh/drisspg/187/orig 2025-09-07T07:36:00.0288222Z * [new branch] gh/drisspg/188/base -> origin/gh/drisspg/188/base 2025-09-07T07:36:00.0288726Z * [new branch] gh/drisspg/188/head -> origin/gh/drisspg/188/head 2025-09-07T07:36:00.0289230Z * [new branch] gh/drisspg/188/orig -> origin/gh/drisspg/188/orig 2025-09-07T07:36:00.0290357Z * [new branch] gh/drisspg/189/base -> origin/gh/drisspg/189/base 2025-09-07T07:36:00.0290733Z * [new branch] gh/drisspg/189/head -> origin/gh/drisspg/189/head 2025-09-07T07:36:00.0291274Z * [new branch] gh/drisspg/189/orig -> origin/gh/drisspg/189/orig 2025-09-07T07:36:00.0292113Z * [new branch] gh/drisspg/190/base -> origin/gh/drisspg/190/base 2025-09-07T07:36:00.0292558Z * [new branch] gh/drisspg/190/head -> origin/gh/drisspg/190/head 2025-09-07T07:36:00.0293276Z * [new branch] gh/drisspg/190/orig -> origin/gh/drisspg/190/orig 2025-09-07T07:36:00.0294011Z * [new branch] gh/drisspg/191/base -> origin/gh/drisspg/191/base 2025-09-07T07:36:00.0294469Z * [new branch] gh/drisspg/191/head -> origin/gh/drisspg/191/head 2025-09-07T07:36:00.0295020Z * [new branch] gh/drisspg/191/orig -> origin/gh/drisspg/191/orig 2025-09-07T07:36:00.0295882Z * [new branch] gh/drisspg/192/base -> origin/gh/drisspg/192/base 2025-09-07T07:36:00.0296321Z * [new branch] gh/drisspg/192/head -> origin/gh/drisspg/192/head 2025-09-07T07:36:00.0296828Z * [new branch] gh/drisspg/192/orig -> origin/gh/drisspg/192/orig 2025-09-07T07:36:00.0297666Z * [new branch] gh/drisspg/193/base -> origin/gh/drisspg/193/base 2025-09-07T07:36:00.0298135Z * [new branch] gh/drisspg/193/head -> origin/gh/drisspg/193/head 2025-09-07T07:36:00.0298737Z * [new branch] gh/drisspg/193/orig -> origin/gh/drisspg/193/orig 2025-09-07T07:36:00.0299504Z * [new branch] gh/drisspg/194/base -> origin/gh/drisspg/194/base 2025-09-07T07:36:00.0299958Z * [new branch] gh/drisspg/194/head -> origin/gh/drisspg/194/head 2025-09-07T07:36:00.0300493Z * [new branch] gh/drisspg/194/orig -> origin/gh/drisspg/194/orig 2025-09-07T07:36:00.0301305Z * [new branch] gh/drisspg/195/base -> origin/gh/drisspg/195/base 2025-09-07T07:36:00.0301843Z * [new branch] gh/drisspg/195/head -> origin/gh/drisspg/195/head 2025-09-07T07:36:00.0302370Z * [new branch] gh/drisspg/195/orig -> origin/gh/drisspg/195/orig 2025-09-07T07:36:00.0303235Z * [new branch] gh/drisspg/196/base -> origin/gh/drisspg/196/base 2025-09-07T07:36:00.0303682Z * [new branch] gh/drisspg/196/head -> origin/gh/drisspg/196/head 2025-09-07T07:36:00.0304247Z * [new branch] gh/drisspg/196/orig -> origin/gh/drisspg/196/orig 2025-09-07T07:36:00.0305056Z * [new branch] gh/drisspg/197/base -> origin/gh/drisspg/197/base 2025-09-07T07:36:00.0305493Z * [new branch] gh/drisspg/197/head -> origin/gh/drisspg/197/head 2025-09-07T07:36:00.0306011Z * [new branch] gh/drisspg/197/orig -> origin/gh/drisspg/197/orig 2025-09-07T07:36:00.0306805Z * [new branch] gh/drisspg/198/base -> origin/gh/drisspg/198/base 2025-09-07T07:36:00.0307248Z * [new branch] gh/drisspg/198/head -> origin/gh/drisspg/198/head 2025-09-07T07:36:00.0307780Z * [new branch] gh/drisspg/198/orig -> origin/gh/drisspg/198/orig 2025-09-07T07:36:00.0308544Z * [new branch] gh/drisspg/199/base -> origin/gh/drisspg/199/base 2025-09-07T07:36:00.0309017Z * [new branch] gh/drisspg/199/head -> origin/gh/drisspg/199/head 2025-09-07T07:36:00.0309506Z * [new branch] gh/drisspg/199/orig -> origin/gh/drisspg/199/orig 2025-09-07T07:36:00.0310686Z * [new branch] gh/dsjohns2/1/base -> origin/gh/dsjohns2/1/base 2025-09-07T07:36:00.0311199Z * [new branch] gh/dsjohns2/1/head -> origin/gh/dsjohns2/1/head 2025-09-07T07:36:00.0312184Z * [new branch] gh/eellison/784/base -> origin/gh/eellison/784/base 2025-09-07T07:36:00.0312657Z * [new branch] gh/eellison/784/head -> origin/gh/eellison/784/head 2025-09-07T07:36:00.0313214Z * [new branch] gh/eellison/784/orig -> origin/gh/eellison/784/orig 2025-09-07T07:36:00.0314289Z * [new branch] gh/eellison/785/base -> origin/gh/eellison/785/base 2025-09-07T07:36:00.0314675Z * [new branch] gh/eellison/785/head -> origin/gh/eellison/785/head 2025-09-07T07:36:00.0315221Z * [new branch] gh/eellison/785/orig -> origin/gh/eellison/785/orig 2025-09-07T07:36:00.0316043Z * [new branch] gh/eellison/789/base -> origin/gh/eellison/789/base 2025-09-07T07:36:00.0316489Z * [new branch] gh/eellison/789/head -> origin/gh/eellison/789/head 2025-09-07T07:36:00.0317022Z * [new branch] gh/eellison/789/orig -> origin/gh/eellison/789/orig 2025-09-07T07:36:00.0317883Z * [new branch] gh/eellison/800/base -> origin/gh/eellison/800/base 2025-09-07T07:36:00.0318396Z * [new branch] gh/eellison/800/head -> origin/gh/eellison/800/head 2025-09-07T07:36:00.0318932Z * [new branch] gh/eellison/800/orig -> origin/gh/eellison/800/orig 2025-09-07T07:36:00.0319740Z * [new branch] gh/eellison/801/base -> origin/gh/eellison/801/base 2025-09-07T07:36:00.0320192Z * [new branch] gh/eellison/801/head -> origin/gh/eellison/801/head 2025-09-07T07:36:00.0320806Z * [new branch] gh/eellison/801/orig -> origin/gh/eellison/801/orig 2025-09-07T07:36:00.0321660Z * [new branch] gh/eellison/802/base -> origin/gh/eellison/802/base 2025-09-07T07:36:00.0322114Z * [new branch] gh/eellison/802/head -> origin/gh/eellison/802/head 2025-09-07T07:36:00.0322661Z * [new branch] gh/eellison/802/orig -> origin/gh/eellison/802/orig 2025-09-07T07:36:00.0323471Z * [new branch] gh/eellison/805/base -> origin/gh/eellison/805/base 2025-09-07T07:36:00.0323922Z * [new branch] gh/eellison/805/head -> origin/gh/eellison/805/head 2025-09-07T07:36:00.0324450Z * [new branch] gh/eellison/805/orig -> origin/gh/eellison/805/orig 2025-09-07T07:36:00.0325415Z * [new branch] gh/eellison/808/base -> origin/gh/eellison/808/base 2025-09-07T07:36:00.0325894Z * [new branch] gh/eellison/808/head -> origin/gh/eellison/808/head 2025-09-07T07:36:00.0326872Z * [new branch] gh/eellison/808/orig -> origin/gh/eellison/808/orig 2025-09-07T07:36:00.0327643Z * [new branch] gh/eellison/809/base -> origin/gh/eellison/809/base 2025-09-07T07:36:00.0328107Z * [new branch] gh/eellison/809/head -> origin/gh/eellison/809/head 2025-09-07T07:36:00.0328639Z * [new branch] gh/eellison/809/orig -> origin/gh/eellison/809/orig 2025-09-07T07:36:00.0329430Z * [new branch] gh/eellison/813/base -> origin/gh/eellison/813/base 2025-09-07T07:36:00.0329889Z * [new branch] gh/eellison/813/head -> origin/gh/eellison/813/head 2025-09-07T07:36:00.0330613Z * [new branch] gh/eellison/813/orig -> origin/gh/eellison/813/orig 2025-09-07T07:36:00.0331357Z * [new branch] gh/eellison/814/base -> origin/gh/eellison/814/base 2025-09-07T07:36:00.0331853Z * [new branch] gh/eellison/814/head -> origin/gh/eellison/814/head 2025-09-07T07:36:00.0332390Z * [new branch] gh/eellison/814/orig -> origin/gh/eellison/814/orig 2025-09-07T07:36:00.0333637Z * [new branch] gh/eellison/815/base -> origin/gh/eellison/815/base 2025-09-07T07:36:00.0334061Z * [new branch] gh/eellison/815/head -> origin/gh/eellison/815/head 2025-09-07T07:36:00.0334579Z * [new branch] gh/eellison/815/orig -> origin/gh/eellison/815/orig 2025-09-07T07:36:00.0335383Z * [new branch] gh/eellison/816/base -> origin/gh/eellison/816/base 2025-09-07T07:36:00.0335855Z * [new branch] gh/eellison/816/head -> origin/gh/eellison/816/head 2025-09-07T07:36:00.0336439Z * [new branch] gh/eellison/816/orig -> origin/gh/eellison/816/orig 2025-09-07T07:36:00.0337255Z * [new branch] gh/eellison/817/base -> origin/gh/eellison/817/base 2025-09-07T07:36:00.0337682Z * [new branch] gh/eellison/817/head -> origin/gh/eellison/817/head 2025-09-07T07:36:00.0338203Z * [new branch] gh/eellison/817/orig -> origin/gh/eellison/817/orig 2025-09-07T07:36:00.0339042Z * [new branch] gh/eellison/818/base -> origin/gh/eellison/818/base 2025-09-07T07:36:00.0339531Z * [new branch] gh/eellison/818/head -> origin/gh/eellison/818/head 2025-09-07T07:36:00.0340257Z * [new branch] gh/eellison/818/orig -> origin/gh/eellison/818/orig 2025-09-07T07:36:00.0341218Z * [new branch] gh/eellison/819/base -> origin/gh/eellison/819/base 2025-09-07T07:36:00.0341642Z * [new branch] gh/eellison/819/head -> origin/gh/eellison/819/head 2025-09-07T07:36:00.0342176Z * [new branch] gh/eellison/819/orig -> origin/gh/eellison/819/orig 2025-09-07T07:36:00.0343482Z * [new branch] gh/eellison/820/base -> origin/gh/eellison/820/base 2025-09-07T07:36:00.0344205Z * [new branch] gh/eellison/820/head -> origin/gh/eellison/820/head 2025-09-07T07:36:00.0344701Z * [new branch] gh/eellison/820/orig -> origin/gh/eellison/820/orig 2025-09-07T07:36:00.0345258Z * [new branch] gh/eellison/821/base -> origin/gh/eellison/821/base 2025-09-07T07:36:00.0345774Z * [new branch] gh/eellison/821/head -> origin/gh/eellison/821/head 2025-09-07T07:36:00.0346352Z * [new branch] gh/eellison/821/orig -> origin/gh/eellison/821/orig 2025-09-07T07:36:00.0347202Z * [new branch] gh/eellison/822/base -> origin/gh/eellison/822/base 2025-09-07T07:36:00.0347614Z * [new branch] gh/eellison/822/head -> origin/gh/eellison/822/head 2025-09-07T07:36:00.0348136Z * [new branch] gh/eellison/822/orig -> origin/gh/eellison/822/orig 2025-09-07T07:36:00.0349079Z * [new branch] gh/eellison/823/base -> origin/gh/eellison/823/base 2025-09-07T07:36:00.0349660Z * [new branch] gh/eellison/823/head -> origin/gh/eellison/823/head 2025-09-07T07:36:00.0350174Z * [new branch] gh/eellison/823/orig -> origin/gh/eellison/823/orig 2025-09-07T07:36:00.0351227Z * [new branch] gh/etaf/132/base -> origin/gh/etaf/132/base 2025-09-07T07:36:00.0351678Z * [new branch] gh/etaf/132/head -> origin/gh/etaf/132/head 2025-09-07T07:36:00.0352202Z * [new branch] gh/etaf/132/orig -> origin/gh/etaf/132/orig 2025-09-07T07:36:00.0353041Z * [new branch] gh/etaf/138/base -> origin/gh/etaf/138/base 2025-09-07T07:36:00.0353451Z * [new branch] gh/etaf/138/head -> origin/gh/etaf/138/head 2025-09-07T07:36:00.0353983Z * [new branch] gh/etaf/138/orig -> origin/gh/etaf/138/orig 2025-09-07T07:36:00.0354855Z * [new branch] gh/etaf/140/base -> origin/gh/etaf/140/base 2025-09-07T07:36:00.0355264Z * [new branch] gh/etaf/140/head -> origin/gh/etaf/140/head 2025-09-07T07:36:00.0355971Z * [new branch] gh/etaf/140/orig -> origin/gh/etaf/140/orig 2025-09-07T07:36:00.0356799Z * [new branch] gh/etaf/143/base -> origin/gh/etaf/143/base 2025-09-07T07:36:00.0357091Z * [new branch] gh/etaf/143/head -> origin/gh/etaf/143/head 2025-09-07T07:36:00.0357641Z * [new branch] gh/etaf/143/orig -> origin/gh/etaf/143/orig 2025-09-07T07:36:00.0358549Z * [new branch] gh/etaf/147/base -> origin/gh/etaf/147/base 2025-09-07T07:36:00.0359121Z * [new branch] gh/etaf/147/head -> origin/gh/etaf/147/head 2025-09-07T07:36:00.0360024Z * [new branch] gh/etaf/151/base -> origin/gh/etaf/151/base 2025-09-07T07:36:00.0360590Z * [new branch] gh/etaf/151/head -> origin/gh/etaf/151/head 2025-09-07T07:36:00.0361155Z * [new branch] gh/etaf/151/orig -> origin/gh/etaf/151/orig 2025-09-07T07:36:00.0362119Z * [new branch] gh/etaf/152/base -> origin/gh/etaf/152/base 2025-09-07T07:36:00.0362590Z * [new branch] gh/etaf/152/head -> origin/gh/etaf/152/head 2025-09-07T07:36:00.0363125Z * [new branch] gh/etaf/152/orig -> origin/gh/etaf/152/orig 2025-09-07T07:36:00.0364074Z * [new branch] gh/etaf/153/base -> origin/gh/etaf/153/base 2025-09-07T07:36:00.0364554Z * [new branch] gh/etaf/153/head -> origin/gh/etaf/153/head 2025-09-07T07:36:00.0365092Z * [new branch] gh/etaf/153/orig -> origin/gh/etaf/153/orig 2025-09-07T07:36:00.0366083Z * [new branch] gh/etaf/154/base -> origin/gh/etaf/154/base 2025-09-07T07:36:00.0366540Z * [new branch] gh/etaf/154/head -> origin/gh/etaf/154/head 2025-09-07T07:36:00.0367060Z * [new branch] gh/etaf/154/orig -> origin/gh/etaf/154/orig 2025-09-07T07:36:00.0368009Z * [new branch] gh/etaf/155/base -> origin/gh/etaf/155/base 2025-09-07T07:36:00.0368483Z * [new branch] gh/etaf/155/head -> origin/gh/etaf/155/head 2025-09-07T07:36:00.0369160Z * [new branch] gh/etaf/155/orig -> origin/gh/etaf/155/orig 2025-09-07T07:36:00.0369971Z * [new branch] gh/etaf/156/base -> origin/gh/etaf/156/base 2025-09-07T07:36:00.0370360Z * [new branch] gh/etaf/156/head -> origin/gh/etaf/156/head 2025-09-07T07:36:00.0370920Z * [new branch] gh/etaf/156/orig -> origin/gh/etaf/156/orig 2025-09-07T07:36:00.0372015Z * [new branch] gh/etaf/157/base -> origin/gh/etaf/157/base 2025-09-07T07:36:00.0372916Z * [new branch] gh/etaf/157/head -> origin/gh/etaf/157/head 2025-09-07T07:36:00.0373365Z * [new branch] gh/etaf/157/orig -> origin/gh/etaf/157/orig 2025-09-07T07:36:00.0374163Z * [new branch] gh/etaf/158/base -> origin/gh/etaf/158/base 2025-09-07T07:36:00.0374646Z * [new branch] gh/etaf/158/head -> origin/gh/etaf/158/head 2025-09-07T07:36:00.0375188Z * [new branch] gh/etaf/158/orig -> origin/gh/etaf/158/orig 2025-09-07T07:36:00.0376095Z * [new branch] gh/etaf/159/base -> origin/gh/etaf/159/base 2025-09-07T07:36:00.0376536Z * [new branch] gh/etaf/159/head -> origin/gh/etaf/159/head 2025-09-07T07:36:00.0377054Z * [new branch] gh/etaf/159/orig -> origin/gh/etaf/159/orig 2025-09-07T07:36:00.0377968Z * [new branch] gh/etaf/160/base -> origin/gh/etaf/160/base 2025-09-07T07:36:00.0378475Z * [new branch] gh/etaf/160/head -> origin/gh/etaf/160/head 2025-09-07T07:36:00.0379231Z * [new branch] gh/etaf/160/orig -> origin/gh/etaf/160/orig 2025-09-07T07:36:00.0380151Z * [new branch] gh/etaf/161/base -> origin/gh/etaf/161/base 2025-09-07T07:36:00.0380721Z * [new branch] gh/etaf/161/head -> origin/gh/etaf/161/head 2025-09-07T07:36:00.0381218Z * [new branch] gh/etaf/161/orig -> origin/gh/etaf/161/orig 2025-09-07T07:36:00.0382310Z * [new branch] gh/etaf/162/base -> origin/gh/etaf/162/base 2025-09-07T07:36:00.0382744Z * [new branch] gh/etaf/162/head -> origin/gh/etaf/162/head 2025-09-07T07:36:00.0383249Z * [new branch] gh/etaf/162/orig -> origin/gh/etaf/162/orig 2025-09-07T07:36:00.0384128Z * [new branch] gh/etaf/163/base -> origin/gh/etaf/163/base 2025-09-07T07:36:00.0384535Z * [new branch] gh/etaf/163/head -> origin/gh/etaf/163/head 2025-09-07T07:36:00.0385055Z * [new branch] gh/etaf/163/orig -> origin/gh/etaf/163/orig 2025-09-07T07:36:00.0385980Z * [new branch] gh/etaf/164/base -> origin/gh/etaf/164/base 2025-09-07T07:36:00.0386444Z * [new branch] gh/etaf/164/head -> origin/gh/etaf/164/head 2025-09-07T07:36:00.0387002Z * [new branch] gh/etaf/164/orig -> origin/gh/etaf/164/orig 2025-09-07T07:36:00.0387950Z * [new branch] gh/etaf/165/base -> origin/gh/etaf/165/base 2025-09-07T07:36:00.0388380Z * [new branch] gh/etaf/165/orig -> origin/gh/etaf/165/orig 2025-09-07T07:36:00.0389366Z * [new branch] gh/etaf/166/base -> origin/gh/etaf/166/base 2025-09-07T07:36:00.0390296Z * [new branch] gh/etaf/166/head -> origin/gh/etaf/166/head 2025-09-07T07:36:00.0390760Z * [new branch] gh/etaf/166/orig -> origin/gh/etaf/166/orig 2025-09-07T07:36:00.0391712Z * [new branch] gh/etaf/167/base -> origin/gh/etaf/167/base 2025-09-07T07:36:00.0392180Z * [new branch] gh/etaf/167/head -> origin/gh/etaf/167/head 2025-09-07T07:36:00.0392728Z * [new branch] gh/etaf/167/orig -> origin/gh/etaf/167/orig 2025-09-07T07:36:00.0393607Z * [new branch] gh/etaf/168/base -> origin/gh/etaf/168/base 2025-09-07T07:36:00.0394113Z * [new branch] gh/etaf/168/head -> origin/gh/etaf/168/head 2025-09-07T07:36:00.0394651Z * [new branch] gh/etaf/168/orig -> origin/gh/etaf/168/orig 2025-09-07T07:36:00.0395671Z * [new branch] gh/etaf/169/base -> origin/gh/etaf/169/base 2025-09-07T07:36:00.0396119Z * [new branch] gh/etaf/169/head -> origin/gh/etaf/169/head 2025-09-07T07:36:00.0396654Z * [new branch] gh/etaf/169/orig -> origin/gh/etaf/169/orig 2025-09-07T07:36:00.0397684Z * [new branch] gh/exclamaforte/1/base -> origin/gh/exclamaforte/1/base 2025-09-07T07:36:00.0398114Z * [new branch] gh/exclamaforte/1/head -> origin/gh/exclamaforte/1/head 2025-09-07T07:36:00.0398997Z * [new branch] gh/exclamaforte/2/base -> origin/gh/exclamaforte/2/base 2025-09-07T07:36:00.0399381Z * [new branch] gh/exclamaforte/2/head -> origin/gh/exclamaforte/2/head 2025-09-07T07:36:00.0400254Z * [new branch] gh/exclamaforte/3/base -> origin/gh/exclamaforte/3/base 2025-09-07T07:36:00.0400700Z * [new branch] gh/exclamaforte/3/head -> origin/gh/exclamaforte/3/head 2025-09-07T07:36:00.0401530Z * [new branch] gh/exclamaforte/4/base -> origin/gh/exclamaforte/4/base 2025-09-07T07:36:00.0401935Z * [new branch] gh/exclamaforte/4/head -> origin/gh/exclamaforte/4/head 2025-09-07T07:36:00.0403159Z * [new branch] gh/ezyang/2374/base -> origin/gh/ezyang/2374/base 2025-09-07T07:36:00.0403622Z * [new branch] gh/ezyang/2374/head -> origin/gh/ezyang/2374/head 2025-09-07T07:36:00.0404208Z * [new branch] gh/ezyang/2374/orig -> origin/gh/ezyang/2374/orig 2025-09-07T07:36:00.0405003Z * [new branch] gh/ezyang/2973/base -> origin/gh/ezyang/2973/base 2025-09-07T07:36:00.0405436Z * [new branch] gh/ezyang/2973/head -> origin/gh/ezyang/2973/head 2025-09-07T07:36:00.0406004Z * [new branch] gh/ezyang/2973/orig -> origin/gh/ezyang/2973/orig 2025-09-07T07:36:00.0406846Z * [new branch] gh/ezyang/2974/base -> origin/gh/ezyang/2974/base 2025-09-07T07:36:00.0407287Z * [new branch] gh/ezyang/2974/head -> origin/gh/ezyang/2974/head 2025-09-07T07:36:00.0407992Z * [new branch] gh/ezyang/2974/orig -> origin/gh/ezyang/2974/orig 2025-09-07T07:36:00.0408829Z * [new branch] gh/ezyang/3074/base -> origin/gh/ezyang/3074/base 2025-09-07T07:36:00.0409250Z * [new branch] gh/ezyang/3074/head -> origin/gh/ezyang/3074/head 2025-09-07T07:36:00.0410058Z * [new branch] gh/ezyang/3074/orig -> origin/gh/ezyang/3074/orig 2025-09-07T07:36:00.0410660Z * [new branch] gh/ezyang/3088/base -> origin/gh/ezyang/3088/base 2025-09-07T07:36:00.0411183Z * [new branch] gh/ezyang/3088/head -> origin/gh/ezyang/3088/head 2025-09-07T07:36:00.0411721Z * [new branch] gh/ezyang/3088/orig -> origin/gh/ezyang/3088/orig 2025-09-07T07:36:00.0412604Z * [new branch] gh/ezyang/3092/base -> origin/gh/ezyang/3092/base 2025-09-07T07:36:00.0413063Z * [new branch] gh/ezyang/3092/head -> origin/gh/ezyang/3092/head 2025-09-07T07:36:00.0413608Z * [new branch] gh/ezyang/3092/orig -> origin/gh/ezyang/3092/orig 2025-09-07T07:36:00.0414486Z * [new branch] gh/ezyang/3103/base -> origin/gh/ezyang/3103/base 2025-09-07T07:36:00.0414823Z * [new branch] gh/ezyang/3103/head -> origin/gh/ezyang/3103/head 2025-09-07T07:36:00.0415342Z * [new branch] gh/ezyang/3103/orig -> origin/gh/ezyang/3103/orig 2025-09-07T07:36:00.0416149Z * [new branch] gh/ezyang/3105/base -> origin/gh/ezyang/3105/base 2025-09-07T07:36:00.0416597Z * [new branch] gh/ezyang/3105/head -> origin/gh/ezyang/3105/head 2025-09-07T07:36:00.0417195Z * [new branch] gh/ezyang/3105/orig -> origin/gh/ezyang/3105/orig 2025-09-07T07:36:00.0418201Z * [new branch] gh/ezyang/3114/base -> origin/gh/ezyang/3114/base 2025-09-07T07:36:00.0418692Z * [new branch] gh/ezyang/3114/head -> origin/gh/ezyang/3114/head 2025-09-07T07:36:00.0419671Z * [new branch] gh/ezyang/3114/orig -> origin/gh/ezyang/3114/orig 2025-09-07T07:36:00.0420424Z * [new branch] gh/ezyang/3116/base -> origin/gh/ezyang/3116/base 2025-09-07T07:36:00.0420856Z * [new branch] gh/ezyang/3116/head -> origin/gh/ezyang/3116/head 2025-09-07T07:36:00.0421390Z * [new branch] gh/ezyang/3116/orig -> origin/gh/ezyang/3116/orig 2025-09-07T07:36:00.0422240Z * [new branch] gh/ezyang/3120/base -> origin/gh/ezyang/3120/base 2025-09-07T07:36:00.0422625Z * [new branch] gh/ezyang/3120/head -> origin/gh/ezyang/3120/head 2025-09-07T07:36:00.0423139Z * [new branch] gh/ezyang/3120/orig -> origin/gh/ezyang/3120/orig 2025-09-07T07:36:00.0423966Z * [new branch] gh/ezyang/3122/base -> origin/gh/ezyang/3122/base 2025-09-07T07:36:00.0424391Z * [new branch] gh/ezyang/3122/head -> origin/gh/ezyang/3122/head 2025-09-07T07:36:00.0424993Z * [new branch] gh/ezyang/3122/orig -> origin/gh/ezyang/3122/orig 2025-09-07T07:36:00.0425827Z * [new branch] gh/ezyang/3123/base -> origin/gh/ezyang/3123/base 2025-09-07T07:36:00.0426232Z * [new branch] gh/ezyang/3123/head -> origin/gh/ezyang/3123/head 2025-09-07T07:36:00.0426963Z * [new branch] gh/ezyang/3123/orig -> origin/gh/ezyang/3123/orig 2025-09-07T07:36:00.0427543Z * [new branch] gh/ezyang/3125/base -> origin/gh/ezyang/3125/base 2025-09-07T07:36:00.0428078Z * [new branch] gh/ezyang/3125/head -> origin/gh/ezyang/3125/head 2025-09-07T07:36:00.0428587Z * [new branch] gh/ezyang/3125/orig -> origin/gh/ezyang/3125/orig 2025-09-07T07:36:00.0429386Z * [new branch] gh/ezyang/3126/base -> origin/gh/ezyang/3126/base 2025-09-07T07:36:00.0429805Z * [new branch] gh/ezyang/3126/head -> origin/gh/ezyang/3126/head 2025-09-07T07:36:00.0430344Z * [new branch] gh/ezyang/3126/orig -> origin/gh/ezyang/3126/orig 2025-09-07T07:36:00.0431459Z * [new branch] gh/ezyang/3127/base -> origin/gh/ezyang/3127/base 2025-09-07T07:36:00.0431904Z * [new branch] gh/ezyang/3127/head -> origin/gh/ezyang/3127/head 2025-09-07T07:36:00.0432678Z * [new branch] gh/ezyang/3127/orig -> origin/gh/ezyang/3127/orig 2025-09-07T07:36:00.0433278Z * [new branch] gh/ezyang/3128/base -> origin/gh/ezyang/3128/base 2025-09-07T07:36:00.0433801Z * [new branch] gh/ezyang/3128/head -> origin/gh/ezyang/3128/head 2025-09-07T07:36:00.0434723Z * [new branch] gh/ezyang/3128/orig -> origin/gh/ezyang/3128/orig 2025-09-07T07:36:00.0435452Z * [new branch] gh/ezyang/3129/base -> origin/gh/ezyang/3129/base 2025-09-07T07:36:00.0435884Z * [new branch] gh/ezyang/3129/head -> origin/gh/ezyang/3129/head 2025-09-07T07:36:00.0436414Z * [new branch] gh/ezyang/3129/orig -> origin/gh/ezyang/3129/orig 2025-09-07T07:36:00.0437408Z * [new branch] gh/ezyang/3130/base -> origin/gh/ezyang/3130/base 2025-09-07T07:36:00.0437854Z * [new branch] gh/ezyang/3130/head -> origin/gh/ezyang/3130/head 2025-09-07T07:36:00.0438394Z * [new branch] gh/ezyang/3130/orig -> origin/gh/ezyang/3130/orig 2025-09-07T07:36:00.0439212Z * [new branch] gh/ezyang/3131/base -> origin/gh/ezyang/3131/base 2025-09-07T07:36:00.0439703Z * [new branch] gh/ezyang/3131/head -> origin/gh/ezyang/3131/head 2025-09-07T07:36:00.0440296Z * [new branch] gh/ezyang/3131/orig -> origin/gh/ezyang/3131/orig 2025-09-07T07:36:00.0441109Z * [new branch] gh/ezyang/3132/base -> origin/gh/ezyang/3132/base 2025-09-07T07:36:00.0441562Z * [new branch] gh/ezyang/3132/head -> origin/gh/ezyang/3132/head 2025-09-07T07:36:00.0442090Z * [new branch] gh/ezyang/3132/orig -> origin/gh/ezyang/3132/orig 2025-09-07T07:36:00.0442886Z * [new branch] gh/ezyang/3133/base -> origin/gh/ezyang/3133/base 2025-09-07T07:36:00.0443306Z * [new branch] gh/ezyang/3133/head -> origin/gh/ezyang/3133/head 2025-09-07T07:36:00.0443848Z * [new branch] gh/ezyang/3133/orig -> origin/gh/ezyang/3133/orig 2025-09-07T07:36:00.0444945Z * [new branch] gh/ezyang/3134/base -> origin/gh/ezyang/3134/base 2025-09-07T07:36:00.0445248Z * [new branch] gh/ezyang/3134/head -> origin/gh/ezyang/3134/head 2025-09-07T07:36:00.0445670Z * [new branch] gh/ezyang/3134/orig -> origin/gh/ezyang/3134/orig 2025-09-07T07:36:00.0446653Z * [new branch] gh/ezyang/3135/base -> origin/gh/ezyang/3135/base 2025-09-07T07:36:00.0447084Z * [new branch] gh/ezyang/3135/head -> origin/gh/ezyang/3135/head 2025-09-07T07:36:00.0447813Z * [new branch] gh/ezyang/3135/orig -> origin/gh/ezyang/3135/orig 2025-09-07T07:36:00.0448408Z * [new branch] gh/ezyang/3136/base -> origin/gh/ezyang/3136/base 2025-09-07T07:36:00.0449120Z * [new branch] gh/ezyang/3136/head -> origin/gh/ezyang/3136/head 2025-09-07T07:36:00.0449479Z * [new branch] gh/ezyang/3136/orig -> origin/gh/ezyang/3136/orig 2025-09-07T07:36:00.0450265Z * [new branch] gh/ezyang/3137/base -> origin/gh/ezyang/3137/base 2025-09-07T07:36:00.0450691Z * [new branch] gh/ezyang/3137/head -> origin/gh/ezyang/3137/head 2025-09-07T07:36:00.0451197Z * [new branch] gh/ezyang/3137/orig -> origin/gh/ezyang/3137/orig 2025-09-07T07:36:00.0451955Z * [new branch] gh/ezyang/3138/base -> origin/gh/ezyang/3138/base 2025-09-07T07:36:00.0452380Z * [new branch] gh/ezyang/3138/head -> origin/gh/ezyang/3138/head 2025-09-07T07:36:00.0452993Z * [new branch] gh/ezyang/3138/orig -> origin/gh/ezyang/3138/orig 2025-09-07T07:36:00.0453788Z * [new branch] gh/ezyang/3139/base -> origin/gh/ezyang/3139/base 2025-09-07T07:36:00.0454220Z * [new branch] gh/ezyang/3139/head -> origin/gh/ezyang/3139/head 2025-09-07T07:36:00.0455028Z * [new branch] gh/ezyang/3139/orig -> origin/gh/ezyang/3139/orig 2025-09-07T07:36:00.0455790Z * [new branch] gh/ezyang/3140/base -> origin/gh/ezyang/3140/base 2025-09-07T07:36:00.0456209Z * [new branch] gh/ezyang/3140/head -> origin/gh/ezyang/3140/head 2025-09-07T07:36:00.0456746Z * [new branch] gh/ezyang/3140/orig -> origin/gh/ezyang/3140/orig 2025-09-07T07:36:00.0457549Z * [new branch] gh/ezyang/3141/base -> origin/gh/ezyang/3141/base 2025-09-07T07:36:00.0457975Z * [new branch] gh/ezyang/3141/head -> origin/gh/ezyang/3141/head 2025-09-07T07:36:00.0458544Z * [new branch] gh/ezyang/3141/orig -> origin/gh/ezyang/3141/orig 2025-09-07T07:36:00.0459374Z * [new branch] gh/ezyang/3142/base -> origin/gh/ezyang/3142/base 2025-09-07T07:36:00.0459839Z * [new branch] gh/ezyang/3142/head -> origin/gh/ezyang/3142/head 2025-09-07T07:36:00.0460365Z * [new branch] gh/ezyang/3142/orig -> origin/gh/ezyang/3142/orig 2025-09-07T07:36:00.0461166Z * [new branch] gh/ezyang/3143/base -> origin/gh/ezyang/3143/base 2025-09-07T07:36:00.0461599Z * [new branch] gh/ezyang/3143/head -> origin/gh/ezyang/3143/head 2025-09-07T07:36:00.0462176Z * [new branch] gh/ezyang/3143/orig -> origin/gh/ezyang/3143/orig 2025-09-07T07:36:00.0463170Z * [new branch] gh/fadara01/1/base -> origin/gh/fadara01/1/base 2025-09-07T07:36:00.0464801Z * [new branch] gh/fadara01/1/head -> origin/gh/fadara01/1/head 2025-09-07T07:36:00.0465412Z * [new branch] gh/fadara01/1/orig -> origin/gh/fadara01/1/orig 2025-09-07T07:36:00.0466592Z * [new branch] gh/fduwjj/171/base -> origin/gh/fduwjj/171/base 2025-09-07T07:36:00.0467126Z * [new branch] gh/fduwjj/171/head -> origin/gh/fduwjj/171/head 2025-09-07T07:36:00.0467643Z * [new branch] gh/fduwjj/171/orig -> origin/gh/fduwjj/171/orig 2025-09-07T07:36:00.0468668Z * [new branch] gh/fduwjj/175/base -> origin/gh/fduwjj/175/base 2025-09-07T07:36:00.0470614Z * [new branch] gh/fduwjj/175/head -> origin/gh/fduwjj/175/head 2025-09-07T07:36:00.0471078Z * [new branch] gh/fduwjj/175/orig -> origin/gh/fduwjj/175/orig 2025-09-07T07:36:00.0472034Z * [new branch] gh/fduwjj/176/base -> origin/gh/fduwjj/176/base 2025-09-07T07:36:00.0472461Z * [new branch] gh/fduwjj/176/head -> origin/gh/fduwjj/176/head 2025-09-07T07:36:00.0472990Z * [new branch] gh/fduwjj/176/orig -> origin/gh/fduwjj/176/orig 2025-09-07T07:36:00.0473803Z * [new branch] gh/fduwjj/177/base -> origin/gh/fduwjj/177/base 2025-09-07T07:36:00.0474330Z * [new branch] gh/fduwjj/177/head -> origin/gh/fduwjj/177/head 2025-09-07T07:36:00.0474838Z * [new branch] gh/fduwjj/177/orig -> origin/gh/fduwjj/177/orig 2025-09-07T07:36:00.0475692Z * [new branch] gh/fduwjj/178/base -> origin/gh/fduwjj/178/base 2025-09-07T07:36:00.0476221Z * [new branch] gh/fduwjj/178/head -> origin/gh/fduwjj/178/head 2025-09-07T07:36:00.0476931Z * [new branch] gh/fduwjj/178/orig -> origin/gh/fduwjj/178/orig 2025-09-07T07:36:00.0477647Z * [new branch] gh/fduwjj/179/base -> origin/gh/fduwjj/179/base 2025-09-07T07:36:00.0478107Z * [new branch] gh/fduwjj/179/head -> origin/gh/fduwjj/179/head 2025-09-07T07:36:00.0478644Z * [new branch] gh/fduwjj/179/orig -> origin/gh/fduwjj/179/orig 2025-09-07T07:36:00.0479586Z * [new branch] gh/fduwjj/180/base -> origin/gh/fduwjj/180/base 2025-09-07T07:36:00.0480052Z * [new branch] gh/fduwjj/180/head -> origin/gh/fduwjj/180/head 2025-09-07T07:36:00.0480587Z * [new branch] gh/fduwjj/180/orig -> origin/gh/fduwjj/180/orig 2025-09-07T07:36:00.0481449Z * [new branch] gh/fduwjj/181/base -> origin/gh/fduwjj/181/base 2025-09-07T07:36:00.0481895Z * [new branch] gh/fduwjj/181/head -> origin/gh/fduwjj/181/head 2025-09-07T07:36:00.0482412Z * [new branch] gh/fduwjj/181/orig -> origin/gh/fduwjj/181/orig 2025-09-07T07:36:00.0483221Z * [new branch] gh/fduwjj/182/base -> origin/gh/fduwjj/182/base 2025-09-07T07:36:00.0483660Z * [new branch] gh/fduwjj/182/head -> origin/gh/fduwjj/182/head 2025-09-07T07:36:00.0484165Z * [new branch] gh/fduwjj/182/orig -> origin/gh/fduwjj/182/orig 2025-09-07T07:36:00.0485044Z * [new branch] gh/fduwjj/183/base -> origin/gh/fduwjj/183/base 2025-09-07T07:36:00.0485757Z * [new branch] gh/fduwjj/183/head -> origin/gh/fduwjj/183/head 2025-09-07T07:36:00.0486358Z * [new branch] gh/fduwjj/183/orig -> origin/gh/fduwjj/183/orig 2025-09-07T07:36:00.0487522Z * [new branch] gh/fduwjj/184/base -> origin/gh/fduwjj/184/base 2025-09-07T07:36:00.0523155Z * [new branch] gh/fduwjj/184/head -> origin/gh/fduwjj/184/head 2025-09-07T07:36:00.0523628Z * [new branch] gh/fduwjj/184/orig -> origin/gh/fduwjj/184/orig 2025-09-07T07:36:00.0523944Z * [new branch] gh/fduwjj/185/base -> origin/gh/fduwjj/185/base 2025-09-07T07:36:00.0524248Z * [new branch] gh/fduwjj/185/head -> origin/gh/fduwjj/185/head 2025-09-07T07:36:00.0524547Z * [new branch] gh/fduwjj/185/orig -> origin/gh/fduwjj/185/orig 2025-09-07T07:36:00.0524847Z * [new branch] gh/fduwjj/186/base -> origin/gh/fduwjj/186/base 2025-09-07T07:36:00.0525139Z * [new branch] gh/fduwjj/186/head -> origin/gh/fduwjj/186/head 2025-09-07T07:36:00.0525431Z * [new branch] gh/fduwjj/186/orig -> origin/gh/fduwjj/186/orig 2025-09-07T07:36:00.0525731Z * [new branch] gh/fduwjj/187/base -> origin/gh/fduwjj/187/base 2025-09-07T07:36:00.0526013Z * [new branch] gh/fduwjj/187/head -> origin/gh/fduwjj/187/head 2025-09-07T07:36:00.0526304Z * [new branch] gh/fduwjj/187/orig -> origin/gh/fduwjj/187/orig 2025-09-07T07:36:00.0526596Z * [new branch] gh/fduwjj/188/base -> origin/gh/fduwjj/188/base 2025-09-07T07:36:00.0526887Z * [new branch] gh/fduwjj/188/head -> origin/gh/fduwjj/188/head 2025-09-07T07:36:00.0527175Z * [new branch] gh/fduwjj/188/orig -> origin/gh/fduwjj/188/orig 2025-09-07T07:36:00.0527547Z * [new branch] gh/fduwjj/189/base -> origin/gh/fduwjj/189/base 2025-09-07T07:36:00.0527852Z * [new branch] gh/fduwjj/189/head -> origin/gh/fduwjj/189/head 2025-09-07T07:36:00.0528153Z * [new branch] gh/fduwjj/189/orig -> origin/gh/fduwjj/189/orig 2025-09-07T07:36:00.0528446Z * [new branch] gh/fduwjj/190/base -> origin/gh/fduwjj/190/base 2025-09-07T07:36:00.0528733Z * [new branch] gh/fduwjj/190/head -> origin/gh/fduwjj/190/head 2025-09-07T07:36:00.0529028Z * [new branch] gh/fduwjj/190/orig -> origin/gh/fduwjj/190/orig 2025-09-07T07:36:00.0529394Z * [new branch] gh/fduwjj/191/base -> origin/gh/fduwjj/191/base 2025-09-07T07:36:00.0529683Z * [new branch] gh/fduwjj/191/head -> origin/gh/fduwjj/191/head 2025-09-07T07:36:00.0529981Z * [new branch] gh/fduwjj/191/orig -> origin/gh/fduwjj/191/orig 2025-09-07T07:36:00.0530272Z * [new branch] gh/fegin/306/base -> origin/gh/fegin/306/base 2025-09-07T07:36:00.0530569Z * [new branch] gh/fegin/306/head -> origin/gh/fegin/306/head 2025-09-07T07:36:00.0530852Z * [new branch] gh/fegin/306/orig -> origin/gh/fegin/306/orig 2025-09-07T07:36:00.0531131Z * [new branch] gh/fegin/307/base -> origin/gh/fegin/307/base 2025-09-07T07:36:00.0531412Z * [new branch] gh/fegin/307/head -> origin/gh/fegin/307/head 2025-09-07T07:36:00.0531705Z * [new branch] gh/fegin/307/orig -> origin/gh/fegin/307/orig 2025-09-07T07:36:00.0532005Z * [new branch] gh/fegin/308/base -> origin/gh/fegin/308/base 2025-09-07T07:36:00.0532302Z * [new branch] gh/fegin/308/head -> origin/gh/fegin/308/head 2025-09-07T07:36:00.0532683Z * [new branch] gh/fegin/308/orig -> origin/gh/fegin/308/orig 2025-09-07T07:36:00.0532976Z * [new branch] gh/fegin/309/base -> origin/gh/fegin/309/base 2025-09-07T07:36:00.0533256Z * [new branch] gh/fegin/309/head -> origin/gh/fegin/309/head 2025-09-07T07:36:00.0533544Z * [new branch] gh/fegin/309/orig -> origin/gh/fegin/309/orig 2025-09-07T07:36:00.0533837Z * [new branch] gh/fegin/310/base -> origin/gh/fegin/310/base 2025-09-07T07:36:00.0534130Z * [new branch] gh/fegin/310/head -> origin/gh/fegin/310/head 2025-09-07T07:36:00.0534245Z * [new branch] gh/fegin/310/orig -> origin/gh/fegin/310/orig 2025-09-07T07:36:00.0534359Z * [new branch] gh/fegin/311/base -> origin/gh/fegin/311/base 2025-09-07T07:36:00.0534484Z * [new branch] gh/fegin/311/head -> origin/gh/fegin/311/head 2025-09-07T07:36:00.0534603Z * [new branch] gh/fegin/311/orig -> origin/gh/fegin/311/orig 2025-09-07T07:36:00.0534728Z * [new branch] gh/fegin/312/base -> origin/gh/fegin/312/base 2025-09-07T07:36:00.0534846Z * [new branch] gh/fegin/312/head -> origin/gh/fegin/312/head 2025-09-07T07:36:00.0534968Z * [new branch] gh/fegin/312/orig -> origin/gh/fegin/312/orig 2025-09-07T07:36:00.0535080Z * [new branch] gh/fegin/313/base -> origin/gh/fegin/313/base 2025-09-07T07:36:00.0535195Z * [new branch] gh/fegin/313/head -> origin/gh/fegin/313/head 2025-09-07T07:36:00.0535318Z * [new branch] gh/fegin/313/orig -> origin/gh/fegin/313/orig 2025-09-07T07:36:00.0535439Z * [new branch] gh/fffrog/124/base -> origin/gh/fffrog/124/base 2025-09-07T07:36:00.0535565Z * [new branch] gh/fffrog/124/head -> origin/gh/fffrog/124/head 2025-09-07T07:36:00.0535685Z * [new branch] gh/fffrog/124/orig -> origin/gh/fffrog/124/orig 2025-09-07T07:36:00.0535835Z * [new branch] gh/fffrog/129/base -> origin/gh/fffrog/129/base 2025-09-07T07:36:00.0535953Z * [new branch] gh/fffrog/129/head -> origin/gh/fffrog/129/head 2025-09-07T07:36:00.0536069Z * [new branch] gh/fffrog/129/orig -> origin/gh/fffrog/129/orig 2025-09-07T07:36:00.0536193Z * [new branch] gh/fffrog/130/base -> origin/gh/fffrog/130/base 2025-09-07T07:36:00.0536308Z * [new branch] gh/fffrog/130/head -> origin/gh/fffrog/130/head 2025-09-07T07:36:00.0536447Z * [new branch] gh/fffrog/130/orig -> origin/gh/fffrog/130/orig 2025-09-07T07:36:00.0536564Z * [new branch] gh/fffrog/131/base -> origin/gh/fffrog/131/base 2025-09-07T07:36:00.0536688Z * [new branch] gh/fffrog/131/head -> origin/gh/fffrog/131/head 2025-09-07T07:36:00.0536809Z * [new branch] gh/fffrog/131/orig -> origin/gh/fffrog/131/orig 2025-09-07T07:36:00.0536927Z * [new branch] gh/fffrog/132/base -> origin/gh/fffrog/132/base 2025-09-07T07:36:00.0537051Z * [new branch] gh/fffrog/132/head -> origin/gh/fffrog/132/head 2025-09-07T07:36:00.0537168Z * [new branch] gh/fffrog/132/orig -> origin/gh/fffrog/132/orig 2025-09-07T07:36:00.0537292Z * [new branch] gh/fffrog/133/base -> origin/gh/fffrog/133/base 2025-09-07T07:36:00.0537412Z * [new branch] gh/fffrog/133/head -> origin/gh/fffrog/133/head 2025-09-07T07:36:00.0537529Z * [new branch] gh/fffrog/133/orig -> origin/gh/fffrog/133/orig 2025-09-07T07:36:00.0537653Z * [new branch] gh/fffrog/134/base -> origin/gh/fffrog/134/base 2025-09-07T07:36:00.0537772Z * [new branch] gh/fffrog/134/head -> origin/gh/fffrog/134/head 2025-09-07T07:36:00.0537930Z * [new branch] gh/fffrog/134/orig -> origin/gh/fffrog/134/orig 2025-09-07T07:36:00.0538055Z * [new branch] gh/fffrog/135/base -> origin/gh/fffrog/135/base 2025-09-07T07:36:00.0538181Z * [new branch] gh/fffrog/135/head -> origin/gh/fffrog/135/head 2025-09-07T07:36:00.0538308Z * [new branch] gh/fffrog/135/orig -> origin/gh/fffrog/135/orig 2025-09-07T07:36:00.0538431Z * [new branch] gh/fffrog/136/base -> origin/gh/fffrog/136/base 2025-09-07T07:36:00.0538561Z * [new branch] gh/fffrog/136/head -> origin/gh/fffrog/136/head 2025-09-07T07:36:00.0538684Z * [new branch] gh/fffrog/136/orig -> origin/gh/fffrog/136/orig 2025-09-07T07:36:00.0538811Z * [new branch] gh/fffrog/137/base -> origin/gh/fffrog/137/base 2025-09-07T07:36:00.0538934Z * [new branch] gh/fffrog/137/head -> origin/gh/fffrog/137/head 2025-09-07T07:36:00.0539065Z * [new branch] gh/fffrog/137/orig -> origin/gh/fffrog/137/orig 2025-09-07T07:36:00.0539191Z * [new branch] gh/fffrog/138/base -> origin/gh/fffrog/138/base 2025-09-07T07:36:00.0539309Z * [new branch] gh/fffrog/138/head -> origin/gh/fffrog/138/head 2025-09-07T07:36:00.0539450Z * [new branch] gh/fffrog/138/orig -> origin/gh/fffrog/138/orig 2025-09-07T07:36:00.0539575Z * [new branch] gh/fffrog/139/base -> origin/gh/fffrog/139/base 2025-09-07T07:36:00.0539709Z * [new branch] gh/fffrog/139/head -> origin/gh/fffrog/139/head 2025-09-07T07:36:00.0539898Z * [new branch] gh/fffrog/139/orig -> origin/gh/fffrog/139/orig 2025-09-07T07:36:00.0540051Z * [new branch] gh/fffrog/140/base -> origin/gh/fffrog/140/base 2025-09-07T07:36:00.0540180Z * [new branch] gh/fffrog/140/head -> origin/gh/fffrog/140/head 2025-09-07T07:36:00.0540591Z * [new branch] gh/fffrog/140/orig -> origin/gh/fffrog/140/orig 2025-09-07T07:36:00.0541423Z * [new branch] gh/fffrog/141/base -> origin/gh/fffrog/141/base 2025-09-07T07:36:00.0542053Z * [new branch] gh/fffrog/141/head -> origin/gh/fffrog/141/head 2025-09-07T07:36:00.0542603Z * [new branch] gh/fffrog/141/orig -> origin/gh/fffrog/141/orig 2025-09-07T07:36:00.0543311Z * [new branch] gh/fffrog/142/base -> origin/gh/fffrog/142/base 2025-09-07T07:36:00.0543876Z * [new branch] gh/fffrog/142/head -> origin/gh/fffrog/142/head 2025-09-07T07:36:00.0544279Z * [new branch] gh/fffrog/142/orig -> origin/gh/fffrog/142/orig 2025-09-07T07:36:00.0545148Z * [new branch] gh/fffrog/143/base -> origin/gh/fffrog/143/base 2025-09-07T07:36:00.0545501Z * [new branch] gh/fffrog/143/head -> origin/gh/fffrog/143/head 2025-09-07T07:36:00.0546121Z * [new branch] gh/fffrog/143/orig -> origin/gh/fffrog/143/orig 2025-09-07T07:36:00.0547179Z * [new branch] gh/fffrog/144/base -> origin/gh/fffrog/144/base 2025-09-07T07:36:00.0547655Z * [new branch] gh/fffrog/144/head -> origin/gh/fffrog/144/head 2025-09-07T07:36:00.0548234Z * [new branch] gh/fffrog/144/orig -> origin/gh/fffrog/144/orig 2025-09-07T07:36:00.0549015Z * [new branch] gh/fffrog/145/base -> origin/gh/fffrog/145/base 2025-09-07T07:36:00.0549580Z * [new branch] gh/fffrog/145/head -> origin/gh/fffrog/145/head 2025-09-07T07:36:00.0549994Z * [new branch] gh/fffrog/145/orig -> origin/gh/fffrog/145/orig 2025-09-07T07:36:00.0550997Z * [new branch] gh/fffrog/146/base -> origin/gh/fffrog/146/base 2025-09-07T07:36:00.0551313Z * [new branch] gh/fffrog/146/head -> origin/gh/fffrog/146/head 2025-09-07T07:36:00.0552030Z * [new branch] gh/fffrog/146/orig -> origin/gh/fffrog/146/orig 2025-09-07T07:36:00.0552776Z * [new branch] gh/fffrog/147/base -> origin/gh/fffrog/147/base 2025-09-07T07:36:00.0553321Z * [new branch] gh/fffrog/147/head -> origin/gh/fffrog/147/head 2025-09-07T07:36:00.0553856Z * [new branch] gh/fffrog/147/orig -> origin/gh/fffrog/147/orig 2025-09-07T07:36:00.0554672Z * [new branch] gh/fffrog/148/base -> origin/gh/fffrog/148/base 2025-09-07T07:36:00.0555293Z * [new branch] gh/fffrog/148/head -> origin/gh/fffrog/148/head 2025-09-07T07:36:00.0555608Z * [new branch] gh/fffrog/148/orig -> origin/gh/fffrog/148/orig 2025-09-07T07:36:00.0556450Z * [new branch] gh/fffrog/149/base -> origin/gh/fffrog/149/base 2025-09-07T07:36:00.0556890Z * [new branch] gh/fffrog/149/head -> origin/gh/fffrog/149/head 2025-09-07T07:36:00.0557601Z * [new branch] gh/fffrog/149/orig -> origin/gh/fffrog/149/orig 2025-09-07T07:36:00.0558246Z * [new branch] gh/fffrog/150/base -> origin/gh/fffrog/150/base 2025-09-07T07:36:00.0558794Z * [new branch] gh/fffrog/150/head -> origin/gh/fffrog/150/head 2025-09-07T07:36:00.0559404Z * [new branch] gh/fffrog/150/orig -> origin/gh/fffrog/150/orig 2025-09-07T07:36:00.0560262Z * [new branch] gh/fffrog/151/base -> origin/gh/fffrog/151/base 2025-09-07T07:36:00.0560907Z * [new branch] gh/fffrog/151/head -> origin/gh/fffrog/151/head 2025-09-07T07:36:00.0561274Z * [new branch] gh/fffrog/151/orig -> origin/gh/fffrog/151/orig 2025-09-07T07:36:00.0562159Z * [new branch] gh/fffrog/152/base -> origin/gh/fffrog/152/base 2025-09-07T07:36:00.0562573Z * [new branch] gh/fffrog/152/head -> origin/gh/fffrog/152/head 2025-09-07T07:36:00.0563431Z * [new branch] gh/fffrog/153/base -> origin/gh/fffrog/153/base 2025-09-07T07:36:00.0563795Z * [new branch] gh/fffrog/153/head -> origin/gh/fffrog/153/head 2025-09-07T07:36:00.0564398Z * [new branch] gh/fffrog/153/orig -> origin/gh/fffrog/153/orig 2025-09-07T07:36:00.0565324Z * [new branch] gh/gmagogsfm/1/base -> origin/gh/gmagogsfm/1/base 2025-09-07T07:36:00.0565921Z * [new branch] gh/gmagogsfm/1/head -> origin/gh/gmagogsfm/1/head 2025-09-07T07:36:00.0566295Z * [new branch] gh/gmagogsfm/1/orig -> origin/gh/gmagogsfm/1/orig 2025-09-07T07:36:00.0567047Z * [new branch] gh/gmagogsfm/2/base -> origin/gh/gmagogsfm/2/base 2025-09-07T07:36:00.0567476Z * [new branch] gh/gmagogsfm/2/head -> origin/gh/gmagogsfm/2/head 2025-09-07T07:36:00.0568085Z * [new branch] gh/gmagogsfm/2/orig -> origin/gh/gmagogsfm/2/orig 2025-09-07T07:36:00.0568918Z * [new branch] gh/gmagogsfm/3/base -> origin/gh/gmagogsfm/3/base 2025-09-07T07:36:00.0569573Z * [new branch] gh/gmagogsfm/3/head -> origin/gh/gmagogsfm/3/head 2025-09-07T07:36:00.0570113Z * [new branch] gh/gmagogsfm/3/orig -> origin/gh/gmagogsfm/3/orig 2025-09-07T07:36:00.0571061Z * [new branch] gh/guangyey/134/base -> origin/gh/guangyey/134/base 2025-09-07T07:36:00.0571654Z * [new branch] gh/guangyey/134/head -> origin/gh/guangyey/134/head 2025-09-07T07:36:00.0572064Z * [new branch] gh/guangyey/134/orig -> origin/gh/guangyey/134/orig 2025-09-07T07:36:00.0572912Z * [new branch] gh/guangyey/135/base -> origin/gh/guangyey/135/base 2025-09-07T07:36:00.0573351Z * [new branch] gh/guangyey/135/head -> origin/gh/guangyey/135/head 2025-09-07T07:36:00.0574317Z * [new branch] gh/guangyey/135/orig -> origin/gh/guangyey/135/orig 2025-09-07T07:36:00.0575086Z * [new branch] gh/guangyey/139/base -> origin/gh/guangyey/139/base 2025-09-07T07:36:00.0575529Z * [new branch] gh/guangyey/139/head -> origin/gh/guangyey/139/head 2025-09-07T07:36:00.0576141Z * [new branch] gh/guangyey/139/orig -> origin/gh/guangyey/139/orig 2025-09-07T07:36:00.0577257Z * [new branch] gh/guangyey/140/base -> origin/gh/guangyey/140/base 2025-09-07T07:36:00.0577710Z * [new branch] gh/guangyey/140/head -> origin/gh/guangyey/140/head 2025-09-07T07:36:00.0578294Z * [new branch] gh/guangyey/140/orig -> origin/gh/guangyey/140/orig 2025-09-07T07:36:00.0579016Z * [new branch] gh/guangyey/142/base -> origin/gh/guangyey/142/base 2025-09-07T07:36:00.0579625Z * [new branch] gh/guangyey/142/head -> origin/gh/guangyey/142/head 2025-09-07T07:36:00.0580180Z * [new branch] gh/guangyey/142/orig -> origin/gh/guangyey/142/orig 2025-09-07T07:36:00.0581227Z * [new branch] gh/guangyey/145/base -> origin/gh/guangyey/145/base 2025-09-07T07:36:00.0581388Z * [new branch] gh/guangyey/145/head -> origin/gh/guangyey/145/head 2025-09-07T07:36:00.0581994Z * [new branch] gh/guangyey/145/orig -> origin/gh/guangyey/145/orig 2025-09-07T07:36:00.0582715Z * [new branch] gh/guangyey/153/base -> origin/gh/guangyey/153/base 2025-09-07T07:36:00.0583282Z * [new branch] gh/guangyey/153/head -> origin/gh/guangyey/153/head 2025-09-07T07:36:00.0583673Z * [new branch] gh/guangyey/153/orig -> origin/gh/guangyey/153/orig 2025-09-07T07:36:00.0584561Z * [new branch] gh/guangyey/159/base -> origin/gh/guangyey/159/base 2025-09-07T07:36:00.0585108Z * [new branch] gh/guangyey/159/head -> origin/gh/guangyey/159/head 2025-09-07T07:36:00.0585617Z * [new branch] gh/guangyey/159/orig -> origin/gh/guangyey/159/orig 2025-09-07T07:36:00.0586332Z * [new branch] gh/guangyey/163/base -> origin/gh/guangyey/163/base 2025-09-07T07:36:00.0587169Z * [new branch] gh/guangyey/163/head -> origin/gh/guangyey/163/head 2025-09-07T07:36:00.0587726Z * [new branch] gh/guangyey/163/orig -> origin/gh/guangyey/163/orig 2025-09-07T07:36:00.0588485Z * [new branch] gh/guangyey/168/base -> origin/gh/guangyey/168/base 2025-09-07T07:36:00.0589106Z * [new branch] gh/guangyey/168/head -> origin/gh/guangyey/168/head 2025-09-07T07:36:00.0589670Z * [new branch] gh/guangyey/168/orig -> origin/gh/guangyey/168/orig 2025-09-07T07:36:00.0590390Z * [new branch] gh/guangyey/169/base -> origin/gh/guangyey/169/base 2025-09-07T07:36:00.0590838Z * [new branch] gh/guangyey/169/head -> origin/gh/guangyey/169/head 2025-09-07T07:36:00.0591432Z * [new branch] gh/guangyey/169/orig -> origin/gh/guangyey/169/orig 2025-09-07T07:36:00.0592207Z * [new branch] gh/guangyey/170/base -> origin/gh/guangyey/170/base 2025-09-07T07:36:00.0592649Z * [new branch] gh/guangyey/170/head -> origin/gh/guangyey/170/head 2025-09-07T07:36:00.0593239Z * [new branch] gh/guangyey/170/orig -> origin/gh/guangyey/170/orig 2025-09-07T07:36:00.0594027Z * [new branch] gh/guangyey/171/base -> origin/gh/guangyey/171/base 2025-09-07T07:36:00.0594568Z * [new branch] gh/guangyey/171/head -> origin/gh/guangyey/171/head 2025-09-07T07:36:00.0595138Z * [new branch] gh/guangyey/171/orig -> origin/gh/guangyey/171/orig 2025-09-07T07:36:00.0595821Z * [new branch] gh/guangyey/174/base -> origin/gh/guangyey/174/base 2025-09-07T07:36:00.0596682Z * [new branch] gh/guangyey/174/head -> origin/gh/guangyey/174/head 2025-09-07T07:36:00.0597223Z * [new branch] gh/guangyey/174/orig -> origin/gh/guangyey/174/orig 2025-09-07T07:36:00.0598061Z * [new branch] gh/guangyey/176/base -> origin/gh/guangyey/176/base 2025-09-07T07:36:00.0598500Z * [new branch] gh/guangyey/176/head -> origin/gh/guangyey/176/head 2025-09-07T07:36:00.0599077Z * [new branch] gh/guangyey/176/orig -> origin/gh/guangyey/176/orig 2025-09-07T07:36:00.0599906Z * [new branch] gh/guangyey/178/base -> origin/gh/guangyey/178/base 2025-09-07T07:36:00.0600330Z * [new branch] gh/guangyey/178/head -> origin/gh/guangyey/178/head 2025-09-07T07:36:00.0600955Z * [new branch] gh/guangyey/178/orig -> origin/gh/guangyey/178/orig 2025-09-07T07:36:00.0601997Z * [new branch] gh/guangyey/181/base -> origin/gh/guangyey/181/base 2025-09-07T07:36:00.0602549Z * [new branch] gh/guangyey/181/head -> origin/gh/guangyey/181/head 2025-09-07T07:36:00.0602983Z * [new branch] gh/guangyey/181/orig -> origin/gh/guangyey/181/orig 2025-09-07T07:36:00.0604087Z * [new branch] gh/guangyey/182/base -> origin/gh/guangyey/182/base 2025-09-07T07:36:00.0604674Z * [new branch] gh/guangyey/182/head -> origin/gh/guangyey/182/head 2025-09-07T07:36:00.0605053Z * [new branch] gh/guangyey/182/orig -> origin/gh/guangyey/182/orig 2025-09-07T07:36:00.0605808Z * [new branch] gh/guangyey/183/base -> origin/gh/guangyey/183/base 2025-09-07T07:36:00.0606350Z * [new branch] gh/guangyey/183/head -> origin/gh/guangyey/183/head 2025-09-07T07:36:00.0606969Z * [new branch] gh/guangyey/183/orig -> origin/gh/guangyey/183/orig 2025-09-07T07:36:00.0607899Z * [new branch] gh/guangyey/184/base -> origin/gh/guangyey/184/base 2025-09-07T07:36:00.0608506Z * [new branch] gh/guangyey/184/head -> origin/gh/guangyey/184/head 2025-09-07T07:36:00.0608893Z * [new branch] gh/guangyey/184/orig -> origin/gh/guangyey/184/orig 2025-09-07T07:36:00.0609759Z * [new branch] gh/guangyey/185/base -> origin/gh/guangyey/185/base 2025-09-07T07:36:00.0610178Z * [new branch] gh/guangyey/185/head -> origin/gh/guangyey/185/head 2025-09-07T07:36:00.0610768Z * [new branch] gh/guangyey/185/orig -> origin/gh/guangyey/185/orig 2025-09-07T07:36:00.0611518Z * [new branch] gh/guangyey/186/base -> origin/gh/guangyey/186/base 2025-09-07T07:36:00.0611973Z * [new branch] gh/guangyey/186/head -> origin/gh/guangyey/186/head 2025-09-07T07:36:00.0612521Z * [new branch] gh/guangyey/186/orig -> origin/gh/guangyey/186/orig 2025-09-07T07:36:00.0613274Z * [new branch] gh/guangyey/187/base -> origin/gh/guangyey/187/base 2025-09-07T07:36:00.0613826Z * [new branch] gh/guangyey/187/head -> origin/gh/guangyey/187/head 2025-09-07T07:36:00.0614251Z * [new branch] gh/guangyey/187/orig -> origin/gh/guangyey/187/orig 2025-09-07T07:36:00.0615099Z * [new branch] gh/guangyey/188/base -> origin/gh/guangyey/188/base 2025-09-07T07:36:00.0615642Z * [new branch] gh/guangyey/188/head -> origin/gh/guangyey/188/head 2025-09-07T07:36:00.0616538Z * [new branch] gh/guangyey/188/orig -> origin/gh/guangyey/188/orig 2025-09-07T07:36:00.0617414Z * [new branch] gh/guangyey/189/base -> origin/gh/guangyey/189/base 2025-09-07T07:36:00.0617863Z * [new branch] gh/guangyey/189/head -> origin/gh/guangyey/189/head 2025-09-07T07:36:00.0618441Z * [new branch] gh/guangyey/189/orig -> origin/gh/guangyey/189/orig 2025-09-07T07:36:00.0619211Z * [new branch] gh/guangyey/190/base -> origin/gh/guangyey/190/base 2025-09-07T07:36:00.0619639Z * [new branch] gh/guangyey/190/head -> origin/gh/guangyey/190/head 2025-09-07T07:36:00.0620224Z * [new branch] gh/guangyey/190/orig -> origin/gh/guangyey/190/orig 2025-09-07T07:36:00.0621231Z * [new branch] gh/guangyey/191/base -> origin/gh/guangyey/191/base 2025-09-07T07:36:00.0621664Z * [new branch] gh/guangyey/191/head -> origin/gh/guangyey/191/head 2025-09-07T07:36:00.0622291Z * [new branch] gh/guangyey/191/orig -> origin/gh/guangyey/191/orig 2025-09-07T07:36:00.0623024Z * [new branch] gh/guangyey/192/base -> origin/gh/guangyey/192/base 2025-09-07T07:36:00.0623535Z * [new branch] gh/guangyey/192/head -> origin/gh/guangyey/192/head 2025-09-07T07:36:00.0624169Z * [new branch] gh/guangyey/192/orig -> origin/gh/guangyey/192/orig 2025-09-07T07:36:00.0624912Z * [new branch] gh/guangyey/193/base -> origin/gh/guangyey/193/base 2025-09-07T07:36:00.0625336Z * [new branch] gh/guangyey/193/head -> origin/gh/guangyey/193/head 2025-09-07T07:36:00.0625952Z * [new branch] gh/guangyey/193/orig -> origin/gh/guangyey/193/orig 2025-09-07T07:36:00.0626786Z * [new branch] gh/guangyey/194/base -> origin/gh/guangyey/194/base 2025-09-07T07:36:00.0627342Z * [new branch] gh/guangyey/194/head -> origin/gh/guangyey/194/head 2025-09-07T07:36:00.0627785Z * [new branch] gh/guangyey/194/orig -> origin/gh/guangyey/194/orig 2025-09-07T07:36:00.0628570Z * [new branch] gh/guangyey/195/base -> origin/gh/guangyey/195/base 2025-09-07T07:36:00.0629181Z * [new branch] gh/guangyey/195/head -> origin/gh/guangyey/195/head 2025-09-07T07:36:00.0629605Z * [new branch] gh/guangyey/195/orig -> origin/gh/guangyey/195/orig 2025-09-07T07:36:00.0630456Z * [new branch] gh/guangyey/196/base -> origin/gh/guangyey/196/base 2025-09-07T07:36:00.0630830Z * [new branch] gh/guangyey/196/head -> origin/gh/guangyey/196/head 2025-09-07T07:36:00.0631598Z * [new branch] gh/guangyey/196/orig -> origin/gh/guangyey/196/orig 2025-09-07T07:36:00.0632358Z * [new branch] gh/guangyey/197/base -> origin/gh/guangyey/197/base 2025-09-07T07:36:00.0632905Z * [new branch] gh/guangyey/197/head -> origin/gh/guangyey/197/head 2025-09-07T07:36:00.0633335Z * [new branch] gh/guangyey/197/orig -> origin/gh/guangyey/197/orig 2025-09-07T07:36:00.0634143Z * [new branch] gh/guangyey/198/base -> origin/gh/guangyey/198/base 2025-09-07T07:36:00.0634578Z * [new branch] gh/guangyey/198/head -> origin/gh/guangyey/198/head 2025-09-07T07:36:00.0635166Z * [new branch] gh/guangyey/198/orig -> origin/gh/guangyey/198/orig 2025-09-07T07:36:00.0636004Z * [new branch] gh/guangyey/199/base -> origin/gh/guangyey/199/base 2025-09-07T07:36:00.0636433Z * [new branch] gh/guangyey/199/head -> origin/gh/guangyey/199/head 2025-09-07T07:36:00.0637063Z * [new branch] gh/guangyey/199/orig -> origin/gh/guangyey/199/orig 2025-09-07T07:36:00.0637793Z * [new branch] gh/guangyey/200/base -> origin/gh/guangyey/200/base 2025-09-07T07:36:00.0638342Z * [new branch] gh/guangyey/200/head -> origin/gh/guangyey/200/head 2025-09-07T07:36:00.0638926Z * [new branch] gh/guangyey/200/orig -> origin/gh/guangyey/200/orig 2025-09-07T07:36:00.0639668Z * [new branch] gh/guangyey/201/base -> origin/gh/guangyey/201/base 2025-09-07T07:36:00.0640097Z * [new branch] gh/guangyey/201/head -> origin/gh/guangyey/201/head 2025-09-07T07:36:00.0640679Z * [new branch] gh/guangyey/201/orig -> origin/gh/guangyey/201/orig 2025-09-07T07:36:00.0641411Z * [new branch] gh/guangyey/202/base -> origin/gh/guangyey/202/base 2025-09-07T07:36:00.0642008Z * [new branch] gh/guangyey/202/head -> origin/gh/guangyey/202/head 2025-09-07T07:36:00.0642347Z * [new branch] gh/guangyey/202/orig -> origin/gh/guangyey/202/orig 2025-09-07T07:36:00.0643170Z * [new branch] gh/guangyey/203/base -> origin/gh/guangyey/203/base 2025-09-07T07:36:00.0643711Z * [new branch] gh/guangyey/203/head -> origin/gh/guangyey/203/head 2025-09-07T07:36:00.0644381Z * [new branch] gh/guangyey/203/orig -> origin/gh/guangyey/203/orig 2025-09-07T07:36:00.0645065Z * [new branch] gh/guangyey/204/base -> origin/gh/guangyey/204/base 2025-09-07T07:36:00.0645641Z * [new branch] gh/guangyey/204/head -> origin/gh/guangyey/204/head 2025-09-07T07:36:00.0646197Z * [new branch] gh/guangyey/204/orig -> origin/gh/guangyey/204/orig 2025-09-07T07:36:00.0646913Z * [new branch] gh/guangyey/205/base -> origin/gh/guangyey/205/base 2025-09-07T07:36:00.0647624Z * [new branch] gh/guangyey/205/head -> origin/gh/guangyey/205/head 2025-09-07T07:36:00.0647870Z * [new branch] gh/guangyey/205/orig -> origin/gh/guangyey/205/orig 2025-09-07T07:36:00.0648667Z * [new branch] gh/guangyey/206/base -> origin/gh/guangyey/206/base 2025-09-07T07:36:00.0649098Z * [new branch] gh/guangyey/206/head -> origin/gh/guangyey/206/head 2025-09-07T07:36:00.0649732Z * [new branch] gh/guangyey/206/orig -> origin/gh/guangyey/206/orig 2025-09-07T07:36:00.0650469Z * [new branch] gh/guangyey/207/base -> origin/gh/guangyey/207/base 2025-09-07T07:36:00.0650916Z * [new branch] gh/guangyey/207/head -> origin/gh/guangyey/207/head 2025-09-07T07:36:00.0651543Z * [new branch] gh/guangyey/207/orig -> origin/gh/guangyey/207/orig 2025-09-07T07:36:00.0652291Z * [new branch] gh/guangyey/79/base -> origin/gh/guangyey/79/base 2025-09-07T07:36:00.0652735Z * [new branch] gh/guangyey/79/head -> origin/gh/guangyey/79/head 2025-09-07T07:36:00.0653510Z * [new branch] gh/guangyey/79/orig -> origin/gh/guangyey/79/orig 2025-09-07T07:36:00.0654225Z * [new branch] gh/guangyey/89/base -> origin/gh/guangyey/89/base 2025-09-07T07:36:00.0654665Z * [new branch] gh/guangyey/89/head -> origin/gh/guangyey/89/head 2025-09-07T07:36:00.0655255Z * [new branch] gh/guangyey/89/orig -> origin/gh/guangyey/89/orig 2025-09-07T07:36:00.0656418Z * [new branch] gh/guilhermeleobas/107/base -> origin/gh/guilhermeleobas/107/base 2025-09-07T07:36:00.0656710Z * [new branch] gh/guilhermeleobas/107/head -> origin/gh/guilhermeleobas/107/head 2025-09-07T07:36:00.0657338Z * [new branch] gh/guilhermeleobas/107/orig -> origin/gh/guilhermeleobas/107/orig 2025-09-07T07:36:00.0658017Z * [new branch] gh/guilhermeleobas/108/base -> origin/gh/guilhermeleobas/108/base 2025-09-07T07:36:00.0658637Z * [new branch] gh/guilhermeleobas/108/head -> origin/gh/guilhermeleobas/108/head 2025-09-07T07:36:00.0659050Z * [new branch] gh/guilhermeleobas/108/orig -> origin/gh/guilhermeleobas/108/orig 2025-09-07T07:36:00.0659859Z * [new branch] gh/guilhermeleobas/124/base -> origin/gh/guilhermeleobas/124/base 2025-09-07T07:36:00.0660431Z * [new branch] gh/guilhermeleobas/124/head -> origin/gh/guilhermeleobas/124/head 2025-09-07T07:36:00.0661004Z * [new branch] gh/guilhermeleobas/124/orig -> origin/gh/guilhermeleobas/124/orig 2025-09-07T07:36:00.0661659Z * [new branch] gh/guilhermeleobas/147/base -> origin/gh/guilhermeleobas/147/base 2025-09-07T07:36:00.0662240Z * [new branch] gh/guilhermeleobas/147/head -> origin/gh/guilhermeleobas/147/head 2025-09-07T07:36:00.0662883Z * [new branch] gh/guilhermeleobas/147/orig -> origin/gh/guilhermeleobas/147/orig 2025-09-07T07:36:00.0663610Z * [new branch] gh/guilhermeleobas/150/base -> origin/gh/guilhermeleobas/150/base 2025-09-07T07:36:00.0664083Z * [new branch] gh/guilhermeleobas/150/head -> origin/gh/guilhermeleobas/150/head 2025-09-07T07:36:00.0664675Z * [new branch] gh/guilhermeleobas/150/orig -> origin/gh/guilhermeleobas/150/orig 2025-09-07T07:36:00.0665435Z * [new branch] gh/guilhermeleobas/163/base -> origin/gh/guilhermeleobas/163/base 2025-09-07T07:36:00.0666018Z * [new branch] gh/guilhermeleobas/163/head -> origin/gh/guilhermeleobas/163/head 2025-09-07T07:36:00.0666451Z * [new branch] gh/guilhermeleobas/163/orig -> origin/gh/guilhermeleobas/163/orig 2025-09-07T07:36:00.0667230Z * [new branch] gh/guilhermeleobas/164/base -> origin/gh/guilhermeleobas/164/base 2025-09-07T07:36:00.0667905Z * [new branch] gh/guilhermeleobas/164/head -> origin/gh/guilhermeleobas/164/head 2025-09-07T07:36:00.0668365Z * [new branch] gh/guilhermeleobas/164/orig -> origin/gh/guilhermeleobas/164/orig 2025-09-07T07:36:00.0669175Z * [new branch] gh/guilhermeleobas/165/base -> origin/gh/guilhermeleobas/165/base 2025-09-07T07:36:00.0669619Z * [new branch] gh/guilhermeleobas/165/head -> origin/gh/guilhermeleobas/165/head 2025-09-07T07:36:00.0670221Z * [new branch] gh/guilhermeleobas/165/orig -> origin/gh/guilhermeleobas/165/orig 2025-09-07T07:36:00.0670986Z * [new branch] gh/guilhermeleobas/166/base -> origin/gh/guilhermeleobas/166/base 2025-09-07T07:36:00.0671340Z * [new branch] gh/guilhermeleobas/166/head -> origin/gh/guilhermeleobas/166/head 2025-09-07T07:36:00.0672066Z * [new branch] gh/guilhermeleobas/166/orig -> origin/gh/guilhermeleobas/166/orig 2025-09-07T07:36:00.0672855Z * [new branch] gh/guilhermeleobas/167/base -> origin/gh/guilhermeleobas/167/base 2025-09-07T07:36:00.0673316Z * [new branch] gh/guilhermeleobas/167/head -> origin/gh/guilhermeleobas/167/head 2025-09-07T07:36:00.0673943Z * [new branch] gh/guilhermeleobas/167/orig -> origin/gh/guilhermeleobas/167/orig 2025-09-07T07:36:00.0674680Z * [new branch] gh/guilhermeleobas/168/base -> origin/gh/guilhermeleobas/168/base 2025-09-07T07:36:00.0675303Z * [new branch] gh/guilhermeleobas/168/head -> origin/gh/guilhermeleobas/168/head 2025-09-07T07:36:00.0675741Z * [new branch] gh/guilhermeleobas/168/orig -> origin/gh/guilhermeleobas/168/orig 2025-09-07T07:36:00.0676561Z * [new branch] gh/guilhermeleobas/169/base -> origin/gh/guilhermeleobas/169/base 2025-09-07T07:36:00.0677007Z * [new branch] gh/guilhermeleobas/169/head -> origin/gh/guilhermeleobas/169/head 2025-09-07T07:36:00.0677624Z * [new branch] gh/guilhermeleobas/169/orig -> origin/gh/guilhermeleobas/169/orig 2025-09-07T07:36:00.0678295Z * [new branch] gh/guilhermeleobas/170/base -> origin/gh/guilhermeleobas/170/base 2025-09-07T07:36:00.0678748Z * [new branch] gh/guilhermeleobas/170/head -> origin/gh/guilhermeleobas/170/head 2025-09-07T07:36:00.0679328Z * [new branch] gh/guilhermeleobas/170/orig -> origin/gh/guilhermeleobas/170/orig 2025-09-07T07:36:00.0680093Z * [new branch] gh/guilhermeleobas/171/base -> origin/gh/guilhermeleobas/171/base 2025-09-07T07:36:00.0680562Z * [new branch] gh/guilhermeleobas/171/head -> origin/gh/guilhermeleobas/171/head 2025-09-07T07:36:00.0681281Z * [new branch] gh/guilhermeleobas/171/orig -> origin/gh/guilhermeleobas/171/orig 2025-09-07T07:36:00.0682000Z * [new branch] gh/guilhermeleobas/173/base -> origin/gh/guilhermeleobas/173/base 2025-09-07T07:36:00.0682581Z * [new branch] gh/guilhermeleobas/173/head -> origin/gh/guilhermeleobas/173/head 2025-09-07T07:36:00.0683029Z * [new branch] gh/guilhermeleobas/173/orig -> origin/gh/guilhermeleobas/173/orig 2025-09-07T07:36:00.0683847Z * [new branch] gh/guilhermeleobas/192/base -> origin/gh/guilhermeleobas/192/base 2025-09-07T07:36:00.0684301Z * [new branch] gh/guilhermeleobas/192/head -> origin/gh/guilhermeleobas/192/head 2025-09-07T07:36:00.0684902Z * [new branch] gh/guilhermeleobas/192/orig -> origin/gh/guilhermeleobas/192/orig 2025-09-07T07:36:00.0685865Z * [new branch] gh/guilhermeleobas/193/base -> origin/gh/guilhermeleobas/193/base 2025-09-07T07:36:00.0686423Z * [new branch] gh/guilhermeleobas/193/head -> origin/gh/guilhermeleobas/193/head 2025-09-07T07:36:00.0687464Z * [new branch] gh/guilhermeleobas/193/orig -> origin/gh/guilhermeleobas/193/orig 2025-09-07T07:36:00.0688231Z * [new branch] gh/guilhermeleobas/194/base -> origin/gh/guilhermeleobas/194/base 2025-09-07T07:36:00.0688842Z * [new branch] gh/guilhermeleobas/194/head -> origin/gh/guilhermeleobas/194/head 2025-09-07T07:36:00.0689237Z * [new branch] gh/guilhermeleobas/194/orig -> origin/gh/guilhermeleobas/194/orig 2025-09-07T07:36:00.0690191Z * [new branch] gh/guilhermeleobas/203/base -> origin/gh/guilhermeleobas/203/base 2025-09-07T07:36:00.0690753Z * [new branch] gh/guilhermeleobas/203/head -> origin/gh/guilhermeleobas/203/head 2025-09-07T07:36:00.0691209Z * [new branch] gh/guilhermeleobas/203/orig -> origin/gh/guilhermeleobas/203/orig 2025-09-07T07:36:00.0692004Z * [new branch] gh/guilhermeleobas/204/base -> origin/gh/guilhermeleobas/204/base 2025-09-07T07:36:00.0692412Z * [new branch] gh/guilhermeleobas/204/head -> origin/gh/guilhermeleobas/204/head 2025-09-07T07:36:00.0692980Z * [new branch] gh/guilhermeleobas/204/orig -> origin/gh/guilhermeleobas/204/orig 2025-09-07T07:36:00.0693944Z * [new branch] gh/guilhermeleobas/205/base -> origin/gh/guilhermeleobas/205/base 2025-09-07T07:36:00.0694183Z * [new branch] gh/guilhermeleobas/205/head -> origin/gh/guilhermeleobas/205/head 2025-09-07T07:36:00.0694823Z * [new branch] gh/guilhermeleobas/205/orig -> origin/gh/guilhermeleobas/205/orig 2025-09-07T07:36:00.0695578Z * [new branch] gh/guilhermeleobas/209/base -> origin/gh/guilhermeleobas/209/base 2025-09-07T07:36:00.0696040Z * [new branch] gh/guilhermeleobas/209/head -> origin/gh/guilhermeleobas/209/head 2025-09-07T07:36:00.0696658Z * [new branch] gh/guilhermeleobas/209/orig -> origin/gh/guilhermeleobas/209/orig 2025-09-07T07:36:00.0697401Z * [new branch] gh/guilhermeleobas/210/base -> origin/gh/guilhermeleobas/210/base 2025-09-07T07:36:00.0698014Z * [new branch] gh/guilhermeleobas/210/head -> origin/gh/guilhermeleobas/210/head 2025-09-07T07:36:00.0698470Z * [new branch] gh/guilhermeleobas/210/orig -> origin/gh/guilhermeleobas/210/orig 2025-09-07T07:36:00.0699297Z * [new branch] gh/guilhermeleobas/211/base -> origin/gh/guilhermeleobas/211/base 2025-09-07T07:36:00.0699932Z * [new branch] gh/guilhermeleobas/211/head -> origin/gh/guilhermeleobas/211/head 2025-09-07T07:36:00.0700386Z * [new branch] gh/guilhermeleobas/211/orig -> origin/gh/guilhermeleobas/211/orig 2025-09-07T07:36:00.0701511Z * [new branch] gh/guilhermeleobas/214/base -> origin/gh/guilhermeleobas/214/base 2025-09-07T07:36:00.0701977Z * [new branch] gh/guilhermeleobas/214/head -> origin/gh/guilhermeleobas/214/head 2025-09-07T07:36:00.0702587Z * [new branch] gh/guilhermeleobas/214/orig -> origin/gh/guilhermeleobas/214/orig 2025-09-07T07:36:00.0703335Z * [new branch] gh/guilhermeleobas/215/base -> origin/gh/guilhermeleobas/215/base 2025-09-07T07:36:00.0703768Z * [new branch] gh/guilhermeleobas/215/head -> origin/gh/guilhermeleobas/215/head 2025-09-07T07:36:00.0704395Z * [new branch] gh/guilhermeleobas/215/orig -> origin/gh/guilhermeleobas/215/orig 2025-09-07T07:36:00.0705135Z * [new branch] gh/guilhermeleobas/216/base -> origin/gh/guilhermeleobas/216/base 2025-09-07T07:36:00.0705754Z * [new branch] gh/guilhermeleobas/216/head -> origin/gh/guilhermeleobas/216/head 2025-09-07T07:36:00.0706190Z * [new branch] gh/guilhermeleobas/216/orig -> origin/gh/guilhermeleobas/216/orig 2025-09-07T07:36:00.0707064Z * [new branch] gh/guilhermeleobas/217/base -> origin/gh/guilhermeleobas/217/base 2025-09-07T07:36:00.0707451Z * [new branch] gh/guilhermeleobas/217/head -> origin/gh/guilhermeleobas/217/head 2025-09-07T07:36:00.0708016Z * [new branch] gh/guilhermeleobas/217/orig -> origin/gh/guilhermeleobas/217/orig 2025-09-07T07:36:00.0708748Z * [new branch] gh/guilhermeleobas/219/base -> origin/gh/guilhermeleobas/219/base 2025-09-07T07:36:00.0709394Z * [new branch] gh/guilhermeleobas/219/head -> origin/gh/guilhermeleobas/219/head 2025-09-07T07:36:00.0709973Z * [new branch] gh/guilhermeleobas/219/orig -> origin/gh/guilhermeleobas/219/orig 2025-09-07T07:36:00.0710674Z * [new branch] gh/guilhermeleobas/220/base -> origin/gh/guilhermeleobas/220/base 2025-09-07T07:36:00.0711238Z * [new branch] gh/guilhermeleobas/220/head -> origin/gh/guilhermeleobas/220/head 2025-09-07T07:36:00.0711682Z * [new branch] gh/guilhermeleobas/220/orig -> origin/gh/guilhermeleobas/220/orig 2025-09-07T07:36:00.0712498Z * [new branch] gh/guilhermeleobas/221/base -> origin/gh/guilhermeleobas/221/base 2025-09-07T07:36:00.0713110Z * [new branch] gh/guilhermeleobas/221/head -> origin/gh/guilhermeleobas/221/head 2025-09-07T07:36:00.0713534Z * [new branch] gh/guilhermeleobas/221/orig -> origin/gh/guilhermeleobas/221/orig 2025-09-07T07:36:00.0714441Z * [new branch] gh/guilhermeleobas/222/base -> origin/gh/guilhermeleobas/222/base 2025-09-07T07:36:00.0714744Z * [new branch] gh/guilhermeleobas/222/head -> origin/gh/guilhermeleobas/222/head 2025-09-07T07:36:00.0715362Z * [new branch] gh/guilhermeleobas/222/orig -> origin/gh/guilhermeleobas/222/orig 2025-09-07T07:36:00.0716106Z * [new branch] gh/guilhermeleobas/223/base -> origin/gh/guilhermeleobas/223/base 2025-09-07T07:36:00.0716674Z * [new branch] gh/guilhermeleobas/223/head -> origin/gh/guilhermeleobas/223/head 2025-09-07T07:36:00.0717653Z * [new branch] gh/guilhermeleobas/223/orig -> origin/gh/guilhermeleobas/223/orig 2025-09-07T07:36:00.0718407Z * [new branch] gh/guilhermeleobas/224/base -> origin/gh/guilhermeleobas/224/base 2025-09-07T07:36:00.0718985Z * [new branch] gh/guilhermeleobas/224/head -> origin/gh/guilhermeleobas/224/head 2025-09-07T07:36:00.0719393Z * [new branch] gh/guilhermeleobas/224/orig -> origin/gh/guilhermeleobas/224/orig 2025-09-07T07:36:00.0720212Z * [new branch] gh/guilhermeleobas/225/base -> origin/gh/guilhermeleobas/225/base 2025-09-07T07:36:00.0720834Z * [new branch] gh/guilhermeleobas/225/head -> origin/gh/guilhermeleobas/225/head 2025-09-07T07:36:00.0721289Z * [new branch] gh/guilhermeleobas/225/orig -> origin/gh/guilhermeleobas/225/orig 2025-09-07T07:36:00.0722045Z * [new branch] gh/guilhermeleobas/226/base -> origin/gh/guilhermeleobas/226/base 2025-09-07T07:36:00.0722477Z * [new branch] gh/guilhermeleobas/226/head -> origin/gh/guilhermeleobas/226/head 2025-09-07T07:36:00.0723087Z * [new branch] gh/guilhermeleobas/226/orig -> origin/gh/guilhermeleobas/226/orig 2025-09-07T07:36:00.0723961Z * [new branch] gh/guilhermeleobas/227/base -> origin/gh/guilhermeleobas/227/base 2025-09-07T07:36:00.0724378Z * [new branch] gh/guilhermeleobas/227/head -> origin/gh/guilhermeleobas/227/head 2025-09-07T07:36:00.0725022Z * [new branch] gh/guilhermeleobas/227/orig -> origin/gh/guilhermeleobas/227/orig 2025-09-07T07:36:00.0725810Z * [new branch] gh/guilhermeleobas/228/base -> origin/gh/guilhermeleobas/228/base 2025-09-07T07:36:00.0726234Z * [new branch] gh/guilhermeleobas/228/head -> origin/gh/guilhermeleobas/228/head 2025-09-07T07:36:00.0726698Z * [new branch] gh/guilhermeleobas/228/orig -> origin/gh/guilhermeleobas/228/orig 2025-09-07T07:36:00.0727469Z * [new branch] gh/guilhermeleobas/229/base -> origin/gh/guilhermeleobas/229/base 2025-09-07T07:36:00.0728164Z * [new branch] gh/guilhermeleobas/229/head -> origin/gh/guilhermeleobas/229/head 2025-09-07T07:36:00.0728628Z * [new branch] gh/guilhermeleobas/229/orig -> origin/gh/guilhermeleobas/229/orig 2025-09-07T07:36:00.0729445Z * [new branch] gh/guilhermeleobas/230/base -> origin/gh/guilhermeleobas/230/base 2025-09-07T07:36:00.0729873Z * [new branch] gh/guilhermeleobas/230/head -> origin/gh/guilhermeleobas/230/head 2025-09-07T07:36:00.0730473Z * [new branch] gh/guilhermeleobas/230/orig -> origin/gh/guilhermeleobas/230/orig 2025-09-07T07:36:00.0731246Z * [new branch] gh/guilhermeleobas/231/base -> origin/gh/guilhermeleobas/231/base 2025-09-07T07:36:00.0731664Z * [new branch] gh/guilhermeleobas/231/head -> origin/gh/guilhermeleobas/231/head 2025-09-07T07:36:00.0732229Z * [new branch] gh/guilhermeleobas/231/orig -> origin/gh/guilhermeleobas/231/orig 2025-09-07T07:36:00.0732978Z * [new branch] gh/guilhermeleobas/232/base -> origin/gh/guilhermeleobas/232/base 2025-09-07T07:36:00.0733392Z * [new branch] gh/guilhermeleobas/232/head -> origin/gh/guilhermeleobas/232/head 2025-09-07T07:36:00.0734014Z * [new branch] gh/guilhermeleobas/232/orig -> origin/gh/guilhermeleobas/232/orig 2025-09-07T07:36:00.0734806Z * [new branch] gh/guilhermeleobas/233/base -> origin/gh/guilhermeleobas/233/base 2025-09-07T07:36:00.0735178Z * [new branch] gh/guilhermeleobas/233/head -> origin/gh/guilhermeleobas/233/head 2025-09-07T07:36:00.0735788Z * [new branch] gh/guilhermeleobas/233/orig -> origin/gh/guilhermeleobas/233/orig 2025-09-07T07:36:00.0736660Z * [new branch] gh/guilhermeleobas/234/base -> origin/gh/guilhermeleobas/234/base 2025-09-07T07:36:00.0737129Z * [new branch] gh/guilhermeleobas/234/head -> origin/gh/guilhermeleobas/234/head 2025-09-07T07:36:00.0737738Z * [new branch] gh/guilhermeleobas/234/orig -> origin/gh/guilhermeleobas/234/orig 2025-09-07T07:36:00.0738485Z * [new branch] gh/guilhermeleobas/235/base -> origin/gh/guilhermeleobas/235/base 2025-09-07T07:36:00.0738937Z * [new branch] gh/guilhermeleobas/235/head -> origin/gh/guilhermeleobas/235/head 2025-09-07T07:36:00.0739621Z * [new branch] gh/guilhermeleobas/235/orig -> origin/gh/guilhermeleobas/235/orig 2025-09-07T07:36:00.0740328Z * [new branch] gh/guilhermeleobas/236/base -> origin/gh/guilhermeleobas/236/base 2025-09-07T07:36:00.0740787Z * [new branch] gh/guilhermeleobas/236/head -> origin/gh/guilhermeleobas/236/head 2025-09-07T07:36:00.0741418Z * [new branch] gh/guilhermeleobas/236/orig -> origin/gh/guilhermeleobas/236/orig 2025-09-07T07:36:00.0742199Z * [new branch] gh/guilhermeleobas/237/base -> origin/gh/guilhermeleobas/237/base 2025-09-07T07:36:00.0742656Z * [new branch] gh/guilhermeleobas/237/head -> origin/gh/guilhermeleobas/237/head 2025-09-07T07:36:00.0743278Z * [new branch] gh/guilhermeleobas/237/orig -> origin/gh/guilhermeleobas/237/orig 2025-09-07T07:36:00.0743978Z * [new branch] gh/guilhermeleobas/238/base -> origin/gh/guilhermeleobas/238/base 2025-09-07T07:36:00.0744489Z * [new branch] gh/guilhermeleobas/238/head -> origin/gh/guilhermeleobas/238/head 2025-09-07T07:36:00.0744937Z * [new branch] gh/guilhermeleobas/238/orig -> origin/gh/guilhermeleobas/238/orig 2025-09-07T07:36:00.0745883Z * [new branch] gh/guilhermeleobas/239/base -> origin/gh/guilhermeleobas/239/base 2025-09-07T07:36:00.0746342Z * [new branch] gh/guilhermeleobas/239/head -> origin/gh/guilhermeleobas/239/head 2025-09-07T07:36:00.0746995Z * [new branch] gh/guilhermeleobas/239/orig -> origin/gh/guilhermeleobas/239/orig 2025-09-07T07:36:00.0747780Z * [new branch] gh/guilhermeleobas/240/base -> origin/gh/guilhermeleobas/240/base 2025-09-07T07:36:00.0748209Z * [new branch] gh/guilhermeleobas/240/head -> origin/gh/guilhermeleobas/240/head 2025-09-07T07:36:00.0748826Z * [new branch] gh/guilhermeleobas/240/orig -> origin/gh/guilhermeleobas/240/orig 2025-09-07T07:36:00.0749562Z * [new branch] gh/guilhermeleobas/241/base -> origin/gh/guilhermeleobas/241/base 2025-09-07T07:36:00.0750155Z * [new branch] gh/guilhermeleobas/241/head -> origin/gh/guilhermeleobas/241/head 2025-09-07T07:36:00.0750603Z * [new branch] gh/guilhermeleobas/241/orig -> origin/gh/guilhermeleobas/241/orig 2025-09-07T07:36:00.0751459Z * [new branch] gh/guilhermeleobas/242/base -> origin/gh/guilhermeleobas/242/base 2025-09-07T07:36:00.0751909Z * [new branch] gh/guilhermeleobas/242/head -> origin/gh/guilhermeleobas/242/head 2025-09-07T07:36:00.0752525Z * [new branch] gh/guilhermeleobas/242/orig -> origin/gh/guilhermeleobas/242/orig 2025-09-07T07:36:00.0753179Z * [new branch] gh/guilhermeleobas/243/base -> origin/gh/guilhermeleobas/243/base 2025-09-07T07:36:00.0753634Z * [new branch] gh/guilhermeleobas/243/head -> origin/gh/guilhermeleobas/243/head 2025-09-07T07:36:00.0754209Z * [new branch] gh/guilhermeleobas/243/orig -> origin/gh/guilhermeleobas/243/orig 2025-09-07T07:36:00.0755111Z * [new branch] gh/guilhermeleobas/244/base -> origin/gh/guilhermeleobas/244/base 2025-09-07T07:36:00.0755491Z * [new branch] gh/guilhermeleobas/244/head -> origin/gh/guilhermeleobas/244/head 2025-09-07T07:36:00.0756098Z * [new branch] gh/guilhermeleobas/244/orig -> origin/gh/guilhermeleobas/244/orig 2025-09-07T07:36:00.0756834Z * [new branch] gh/guilhermeleobas/245/base -> origin/gh/guilhermeleobas/245/base 2025-09-07T07:36:00.0757394Z * [new branch] gh/guilhermeleobas/245/head -> origin/gh/guilhermeleobas/245/head 2025-09-07T07:36:00.0757826Z * [new branch] gh/guilhermeleobas/245/orig -> origin/gh/guilhermeleobas/245/orig 2025-09-07T07:36:00.0758681Z * [new branch] gh/guilhermeleobas/73/base -> origin/gh/guilhermeleobas/73/base 2025-09-07T07:36:00.0759299Z * [new branch] gh/guilhermeleobas/73/head -> origin/gh/guilhermeleobas/73/head 2025-09-07T07:36:00.0759689Z * [new branch] gh/guilhermeleobas/73/orig -> origin/gh/guilhermeleobas/73/orig 2025-09-07T07:36:00.0760711Z * [new branch] gh/henrylhtsang/140/base -> origin/gh/henrylhtsang/140/base 2025-09-07T07:36:00.0761319Z * [new branch] gh/henrylhtsang/140/head -> origin/gh/henrylhtsang/140/head 2025-09-07T07:36:00.0761752Z * [new branch] gh/henrylhtsang/140/orig -> origin/gh/henrylhtsang/140/orig 2025-09-07T07:36:00.0762481Z * [new branch] gh/henrylhtsang/141/base -> origin/gh/henrylhtsang/141/base 2025-09-07T07:36:00.0762940Z * [new branch] gh/henrylhtsang/141/head -> origin/gh/henrylhtsang/141/head 2025-09-07T07:36:00.0763543Z * [new branch] gh/henrylhtsang/141/orig -> origin/gh/henrylhtsang/141/orig 2025-09-07T07:36:00.0764723Z * [new branch] gh/henrylhtsang/142/base -> origin/gh/henrylhtsang/142/base 2025-09-07T07:36:00.0765336Z * [new branch] gh/henrylhtsang/142/head -> origin/gh/henrylhtsang/142/head 2025-09-07T07:36:00.0765934Z * [new branch] gh/henrylhtsang/142/orig -> origin/gh/henrylhtsang/142/orig 2025-09-07T07:36:00.0766733Z * [new branch] gh/henrylhtsang/143/base -> origin/gh/henrylhtsang/143/base 2025-09-07T07:36:00.0767113Z * [new branch] gh/henrylhtsang/143/head -> origin/gh/henrylhtsang/143/head 2025-09-07T07:36:00.0767733Z * [new branch] gh/henrylhtsang/143/orig -> origin/gh/henrylhtsang/143/orig 2025-09-07T07:36:00.0768479Z * [new branch] gh/henrylhtsang/144/base -> origin/gh/henrylhtsang/144/base 2025-09-07T07:36:00.0769085Z * [new branch] gh/henrylhtsang/144/head -> origin/gh/henrylhtsang/144/head 2025-09-07T07:36:00.0769670Z * [new branch] gh/henrylhtsang/144/orig -> origin/gh/henrylhtsang/144/orig 2025-09-07T07:36:00.0770387Z * [new branch] gh/henrylhtsang/145/base -> origin/gh/henrylhtsang/145/base 2025-09-07T07:36:00.0770958Z * [new branch] gh/henrylhtsang/145/head -> origin/gh/henrylhtsang/145/head 2025-09-07T07:36:00.0771404Z * [new branch] gh/henrylhtsang/145/orig -> origin/gh/henrylhtsang/145/orig 2025-09-07T07:36:00.0772343Z * [new branch] gh/henrylhtsang/146/base -> origin/gh/henrylhtsang/146/base 2025-09-07T07:36:00.0772797Z * [new branch] gh/henrylhtsang/146/head -> origin/gh/henrylhtsang/146/head 2025-09-07T07:36:00.0773510Z * [new branch] gh/henrylhtsang/146/orig -> origin/gh/henrylhtsang/146/orig 2025-09-07T07:36:00.0774290Z * [new branch] gh/henrylhtsang/147/base -> origin/gh/henrylhtsang/147/base 2025-09-07T07:36:00.0774730Z * [new branch] gh/henrylhtsang/147/head -> origin/gh/henrylhtsang/147/head 2025-09-07T07:36:00.0775327Z * [new branch] gh/henrylhtsang/147/orig -> origin/gh/henrylhtsang/147/orig 2025-09-07T07:36:00.0776230Z * [new branch] gh/henrylhtsang/148/base -> origin/gh/henrylhtsang/148/base 2025-09-07T07:36:00.0776885Z * [new branch] gh/henrylhtsang/148/head -> origin/gh/henrylhtsang/148/head 2025-09-07T07:36:00.0777442Z * [new branch] gh/henrylhtsang/148/orig -> origin/gh/henrylhtsang/148/orig 2025-09-07T07:36:00.0778174Z * [new branch] gh/henrylhtsang/149/base -> origin/gh/henrylhtsang/149/base 2025-09-07T07:36:00.0778741Z * [new branch] gh/henrylhtsang/149/head -> origin/gh/henrylhtsang/149/head 2025-09-07T07:36:00.0779233Z * [new branch] gh/henrylhtsang/149/orig -> origin/gh/henrylhtsang/149/orig 2025-09-07T07:36:00.0780305Z * [new branch] gh/huydhn/1/next -> origin/gh/huydhn/1/next 2025-09-07T07:36:00.0780990Z * [new branch] gh/huydhn/2/next -> origin/gh/huydhn/2/next 2025-09-07T07:36:00.0781714Z * [new branch] gh/huydhn/3/next -> origin/gh/huydhn/3/next 2025-09-07T07:36:00.0782422Z * [new branch] gh/huydhn/4/next -> origin/gh/huydhn/4/next 2025-09-07T07:36:00.0783146Z * [new branch] gh/huydhn/5/next -> origin/gh/huydhn/5/next 2025-09-07T07:36:00.0783978Z * [new branch] gh/huydhn/6/next -> origin/gh/huydhn/6/next 2025-09-07T07:36:00.0785285Z * [new branch] gh/int3/97/base -> origin/gh/int3/97/base 2025-09-07T07:36:00.0785824Z * [new branch] gh/int3/97/head -> origin/gh/int3/97/head 2025-09-07T07:36:00.0787142Z * [new branch] gh/isuruf/101/base -> origin/gh/isuruf/101/base 2025-09-07T07:36:00.0787683Z * [new branch] gh/isuruf/101/head -> origin/gh/isuruf/101/head 2025-09-07T07:36:00.0788681Z * [new branch] gh/isuruf/141/base -> origin/gh/isuruf/141/base 2025-09-07T07:36:00.0789163Z * [new branch] gh/isuruf/141/head -> origin/gh/isuruf/141/head 2025-09-07T07:36:00.0789751Z * [new branch] gh/isuruf/141/orig -> origin/gh/isuruf/141/orig 2025-09-07T07:36:00.0790477Z * [new branch] gh/isuruf/142/base -> origin/gh/isuruf/142/base 2025-09-07T07:36:00.0791097Z * [new branch] gh/isuruf/142/head -> origin/gh/isuruf/142/head 2025-09-07T07:36:00.0791483Z * [new branch] gh/isuruf/142/orig -> origin/gh/isuruf/142/orig 2025-09-07T07:36:00.0792235Z * [new branch] gh/isuruf/143/base -> origin/gh/isuruf/143/base 2025-09-07T07:36:00.0792678Z * [new branch] gh/isuruf/143/head -> origin/gh/isuruf/143/head 2025-09-07T07:36:00.0793272Z * [new branch] gh/isuruf/143/orig -> origin/gh/isuruf/143/orig 2025-09-07T07:36:00.0794039Z * [new branch] gh/isuruf/144/base -> origin/gh/isuruf/144/base 2025-09-07T07:36:00.0794696Z * [new branch] gh/isuruf/144/head -> origin/gh/isuruf/144/head 2025-09-07T07:36:00.0795256Z * [new branch] gh/isuruf/144/orig -> origin/gh/isuruf/144/orig 2025-09-07T07:36:00.0795965Z * [new branch] gh/isuruf/145/base -> origin/gh/isuruf/145/base 2025-09-07T07:36:00.0796402Z * [new branch] gh/isuruf/145/head -> origin/gh/isuruf/145/head 2025-09-07T07:36:00.0797022Z * [new branch] gh/isuruf/145/orig -> origin/gh/isuruf/145/orig 2025-09-07T07:36:00.0797818Z * [new branch] gh/isuruf/146/base -> origin/gh/isuruf/146/base 2025-09-07T07:36:00.0798203Z * [new branch] gh/isuruf/146/head -> origin/gh/isuruf/146/head 2025-09-07T07:36:00.0798807Z * [new branch] gh/isuruf/146/orig -> origin/gh/isuruf/146/orig 2025-09-07T07:36:00.0799520Z * [new branch] gh/isuruf/81/base -> origin/gh/isuruf/81/base 2025-09-07T07:36:00.0799953Z * [new branch] gh/isuruf/81/head -> origin/gh/isuruf/81/head 2025-09-07T07:36:00.0800674Z * [new branch] gh/isuruf/81/orig -> origin/gh/isuruf/81/orig 2025-09-07T07:36:00.0801544Z * [new branch] gh/jamesjwu/150/base -> origin/gh/jamesjwu/150/base 2025-09-07T07:36:00.0801991Z * [new branch] gh/jamesjwu/150/head -> origin/gh/jamesjwu/150/head 2025-09-07T07:36:00.0802607Z * [new branch] gh/jamesjwu/150/orig -> origin/gh/jamesjwu/150/orig 2025-09-07T07:36:00.0803474Z * [new branch] gh/jamesjwu/154/base -> origin/gh/jamesjwu/154/base 2025-09-07T07:36:00.0804064Z * [new branch] gh/jamesjwu/154/head -> origin/gh/jamesjwu/154/head 2025-09-07T07:36:00.0804655Z * [new branch] gh/jamesjwu/154/orig -> origin/gh/jamesjwu/154/orig 2025-09-07T07:36:00.0805298Z * [new branch] gh/jamesjwu/155/base -> origin/gh/jamesjwu/155/base 2025-09-07T07:36:00.0805755Z * [new branch] gh/jamesjwu/155/head -> origin/gh/jamesjwu/155/head 2025-09-07T07:36:00.0806376Z * [new branch] gh/jamesjwu/155/orig -> origin/gh/jamesjwu/155/orig 2025-09-07T07:36:00.0807072Z * [new branch] gh/jamesjwu/159/base -> origin/gh/jamesjwu/159/base 2025-09-07T07:36:00.0807512Z * [new branch] gh/jamesjwu/159/head -> origin/gh/jamesjwu/159/head 2025-09-07T07:36:00.0808221Z * [new branch] gh/jamesjwu/159/orig -> origin/gh/jamesjwu/159/orig 2025-09-07T07:36:00.0809227Z * [new branch] gh/jamesjwu/163/base -> origin/gh/jamesjwu/163/base 2025-09-07T07:36:00.0809797Z * [new branch] gh/jamesjwu/163/head -> origin/gh/jamesjwu/163/head 2025-09-07T07:36:00.0810208Z * [new branch] gh/jamesjwu/163/orig -> origin/gh/jamesjwu/163/orig 2025-09-07T07:36:00.0810981Z * [new branch] gh/jamesjwu/171/base -> origin/gh/jamesjwu/171/base 2025-09-07T07:36:00.0811416Z * [new branch] gh/jamesjwu/171/head -> origin/gh/jamesjwu/171/head 2025-09-07T07:36:00.0811999Z * [new branch] gh/jamesjwu/171/orig -> origin/gh/jamesjwu/171/orig 2025-09-07T07:36:00.0812740Z * [new branch] gh/jamesjwu/176/base -> origin/gh/jamesjwu/176/base 2025-09-07T07:36:00.0813373Z * [new branch] gh/jamesjwu/176/head -> origin/gh/jamesjwu/176/head 2025-09-07T07:36:00.0813873Z * [new branch] gh/jamesjwu/176/orig -> origin/gh/jamesjwu/176/orig 2025-09-07T07:36:00.0814648Z * [new branch] gh/jamesjwu/181/base -> origin/gh/jamesjwu/181/base 2025-09-07T07:36:00.0815094Z * [new branch] gh/jamesjwu/181/head -> origin/gh/jamesjwu/181/head 2025-09-07T07:36:00.0815822Z * [new branch] gh/jamesjwu/181/orig -> origin/gh/jamesjwu/181/orig 2025-09-07T07:36:00.0816577Z * [new branch] gh/jamesjwu/182/base -> origin/gh/jamesjwu/182/base 2025-09-07T07:36:00.0817130Z * [new branch] gh/jamesjwu/182/head -> origin/gh/jamesjwu/182/head 2025-09-07T07:36:00.0817574Z * [new branch] gh/jamesjwu/182/orig -> origin/gh/jamesjwu/182/orig 2025-09-07T07:36:00.0818991Z * [new branch] gh/jamesjwu/183/base -> origin/gh/jamesjwu/183/base 2025-09-07T07:36:00.0819536Z * [new branch] gh/jamesjwu/183/head -> origin/gh/jamesjwu/183/head 2025-09-07T07:36:00.0820129Z * [new branch] gh/jamesjwu/183/orig -> origin/gh/jamesjwu/183/orig 2025-09-07T07:36:00.0820772Z * [new branch] gh/jamesjwu/184/base -> origin/gh/jamesjwu/184/base 2025-09-07T07:36:00.0821324Z * [new branch] gh/jamesjwu/184/head -> origin/gh/jamesjwu/184/head 2025-09-07T07:36:00.0821767Z * [new branch] gh/jamesjwu/184/orig -> origin/gh/jamesjwu/184/orig 2025-09-07T07:36:00.0822683Z * [new branch] gh/jamesjwu/185/base -> origin/gh/jamesjwu/185/base 2025-09-07T07:36:00.0823273Z * [new branch] gh/jamesjwu/185/head -> origin/gh/jamesjwu/185/head 2025-09-07T07:36:00.0823815Z * [new branch] gh/jamesjwu/185/orig -> origin/gh/jamesjwu/185/orig 2025-09-07T07:36:00.0824526Z * [new branch] gh/jamesjwu/186/base -> origin/gh/jamesjwu/186/base 2025-09-07T07:36:00.0824956Z * [new branch] gh/jamesjwu/186/head -> origin/gh/jamesjwu/186/head 2025-09-07T07:36:00.0825568Z * [new branch] gh/jamesjwu/186/orig -> origin/gh/jamesjwu/186/orig 2025-09-07T07:36:00.0826270Z * [new branch] gh/jamesjwu/187/base -> origin/gh/jamesjwu/187/base 2025-09-07T07:36:00.0826708Z * [new branch] gh/jamesjwu/187/head -> origin/gh/jamesjwu/187/head 2025-09-07T07:36:00.0827321Z * [new branch] gh/jamesjwu/187/orig -> origin/gh/jamesjwu/187/orig 2025-09-07T07:36:00.0828055Z * [new branch] gh/jamesjwu/188/base -> origin/gh/jamesjwu/188/base 2025-09-07T07:36:00.0828489Z * [new branch] gh/jamesjwu/188/head -> origin/gh/jamesjwu/188/head 2025-09-07T07:36:00.0829095Z * [new branch] gh/jamesjwu/188/orig -> origin/gh/jamesjwu/188/orig 2025-09-07T07:36:00.0829797Z * [new branch] gh/jamesjwu/189/base -> origin/gh/jamesjwu/189/base 2025-09-07T07:36:00.0830337Z * [new branch] gh/jamesjwu/189/head -> origin/gh/jamesjwu/189/head 2025-09-07T07:36:00.0830924Z * [new branch] gh/jamesjwu/189/orig -> origin/gh/jamesjwu/189/orig 2025-09-07T07:36:00.0832029Z * [new branch] gh/jamesjwu/190/base -> origin/gh/jamesjwu/190/base 2025-09-07T07:36:00.0832464Z * [new branch] gh/jamesjwu/190/head -> origin/gh/jamesjwu/190/head 2025-09-07T07:36:00.0833428Z * [new branch] gh/jamesjwu/190/orig -> origin/gh/jamesjwu/190/orig 2025-09-07T07:36:00.0834287Z * [new branch] gh/jamesjwu/52/base -> origin/gh/jamesjwu/52/base 2025-09-07T07:36:00.0834851Z * [new branch] gh/jamesjwu/52/head -> origin/gh/jamesjwu/52/head 2025-09-07T07:36:00.0835587Z * [new branch] gh/jamesjwu/53/base -> origin/gh/jamesjwu/53/base 2025-09-07T07:36:00.0835946Z * [new branch] gh/jamesjwu/53/head -> origin/gh/jamesjwu/53/head 2025-09-07T07:36:00.0836685Z * [new branch] gh/jamesjwu/54/base -> origin/gh/jamesjwu/54/base 2025-09-07T07:36:00.0837075Z * [new branch] gh/jamesjwu/54/head -> origin/gh/jamesjwu/54/head 2025-09-07T07:36:00.0837875Z * [new branch] gh/jamesjwu/55/base -> origin/gh/jamesjwu/55/base 2025-09-07T07:36:00.0838465Z * [new branch] gh/jamesjwu/55/head -> origin/gh/jamesjwu/55/head 2025-09-07T07:36:00.0839079Z * [new branch] gh/jamesjwu/56/base -> origin/gh/jamesjwu/56/base 2025-09-07T07:36:00.0839490Z * [new branch] gh/jamesjwu/56/head -> origin/gh/jamesjwu/56/head 2025-09-07T07:36:00.0840346Z * [new branch] gh/jamesjwu/57/base -> origin/gh/jamesjwu/57/base 2025-09-07T07:36:00.0840606Z * [new branch] gh/jamesjwu/57/head -> origin/gh/jamesjwu/57/head 2025-09-07T07:36:00.0841470Z * [new branch] gh/jamesjwu/58/base -> origin/gh/jamesjwu/58/base 2025-09-07T07:36:00.0841874Z * [new branch] gh/jamesjwu/58/head -> origin/gh/jamesjwu/58/head 2025-09-07T07:36:00.0842611Z * [new branch] gh/jamesjwu/59/base -> origin/gh/jamesjwu/59/base 2025-09-07T07:36:00.0843039Z * [new branch] gh/jamesjwu/59/head -> origin/gh/jamesjwu/59/head 2025-09-07T07:36:00.0843802Z * [new branch] gh/jamesjwu/60/base -> origin/gh/jamesjwu/60/base 2025-09-07T07:36:00.0844178Z * [new branch] gh/jamesjwu/60/head -> origin/gh/jamesjwu/60/head 2025-09-07T07:36:00.0845043Z * [new branch] gh/jamesjwu/61/base -> origin/gh/jamesjwu/61/base 2025-09-07T07:36:00.0845637Z * [new branch] gh/jamesjwu/61/head -> origin/gh/jamesjwu/61/head 2025-09-07T07:36:00.0846295Z * [new branch] gh/jamesjwu/62/base -> origin/gh/jamesjwu/62/base 2025-09-07T07:36:00.0846689Z * [new branch] gh/jamesjwu/62/head -> origin/gh/jamesjwu/62/head 2025-09-07T07:36:00.0847432Z * [new branch] gh/jamesjwu/63/base -> origin/gh/jamesjwu/63/base 2025-09-07T07:36:00.0847879Z * [new branch] gh/jamesjwu/63/head -> origin/gh/jamesjwu/63/head 2025-09-07T07:36:00.0848808Z * [new branch] gh/jamesjwu/64/base -> origin/gh/jamesjwu/64/base 2025-09-07T07:36:00.0849241Z * [new branch] gh/jamesjwu/64/head -> origin/gh/jamesjwu/64/head 2025-09-07T07:36:00.0850007Z * [new branch] gh/jamesjwu/65/base -> origin/gh/jamesjwu/65/base 2025-09-07T07:36:00.0850599Z * [new branch] gh/jamesjwu/65/head -> origin/gh/jamesjwu/65/head 2025-09-07T07:36:00.0851975Z * [new branch] gh/janeyx99/165/base -> origin/gh/janeyx99/165/base 2025-09-07T07:36:00.0852569Z * [new branch] gh/janeyx99/165/head -> origin/gh/janeyx99/165/head 2025-09-07T07:36:00.0853119Z * [new branch] gh/janeyx99/165/orig -> origin/gh/janeyx99/165/orig 2025-09-07T07:36:00.0853822Z * [new branch] gh/janeyx99/201/base -> origin/gh/janeyx99/201/base 2025-09-07T07:36:00.0854403Z * [new branch] gh/janeyx99/201/head -> origin/gh/janeyx99/201/head 2025-09-07T07:36:00.0854789Z * [new branch] gh/janeyx99/201/orig -> origin/gh/janeyx99/201/orig 2025-09-07T07:36:00.0855855Z * [new branch] gh/janeyx99/225/base -> origin/gh/janeyx99/225/base 2025-09-07T07:36:00.0856286Z * [new branch] gh/janeyx99/225/head -> origin/gh/janeyx99/225/head 2025-09-07T07:36:00.0856949Z * [new branch] gh/janeyx99/225/orig -> origin/gh/janeyx99/225/orig 2025-09-07T07:36:00.0857717Z * [new branch] gh/janeyx99/296/base -> origin/gh/janeyx99/296/base 2025-09-07T07:36:00.0858253Z * [new branch] gh/janeyx99/296/head -> origin/gh/janeyx99/296/head 2025-09-07T07:36:00.0858695Z * [new branch] gh/janeyx99/296/orig -> origin/gh/janeyx99/296/orig 2025-09-07T07:36:00.0859560Z * [new branch] gh/janeyx99/297/base -> origin/gh/janeyx99/297/base 2025-09-07T07:36:00.0860008Z * [new branch] gh/janeyx99/297/head -> origin/gh/janeyx99/297/head 2025-09-07T07:36:00.0860710Z * [new branch] gh/janeyx99/297/orig -> origin/gh/janeyx99/297/orig 2025-09-07T07:36:00.0861466Z * [new branch] gh/janeyx99/298/base -> origin/gh/janeyx99/298/base 2025-09-07T07:36:00.0862003Z * [new branch] gh/janeyx99/298/head -> origin/gh/janeyx99/298/head 2025-09-07T07:36:00.0862438Z * [new branch] gh/janeyx99/298/orig -> origin/gh/janeyx99/298/orig 2025-09-07T07:36:00.0863262Z * [new branch] gh/janeyx99/299/base -> origin/gh/janeyx99/299/base 2025-09-07T07:36:00.0863697Z * [new branch] gh/janeyx99/299/head -> origin/gh/janeyx99/299/head 2025-09-07T07:36:00.0864293Z * [new branch] gh/janeyx99/299/orig -> origin/gh/janeyx99/299/orig 2025-09-07T07:36:00.0865148Z * [new branch] gh/janeyx99/300/base -> origin/gh/janeyx99/300/base 2025-09-07T07:36:00.0865821Z * [new branch] gh/janeyx99/300/head -> origin/gh/janeyx99/300/head 2025-09-07T07:36:00.0866258Z * [new branch] gh/janeyx99/300/orig -> origin/gh/janeyx99/300/orig 2025-09-07T07:36:00.0867190Z * [new branch] gh/janeyx99/301/base -> origin/gh/janeyx99/301/base 2025-09-07T07:36:00.0867639Z * [new branch] gh/janeyx99/301/head -> origin/gh/janeyx99/301/head 2025-09-07T07:36:00.0868235Z * [new branch] gh/janeyx99/301/orig -> origin/gh/janeyx99/301/orig 2025-09-07T07:36:00.0873121Z * [new branch] gh/janeyx99/302/base -> origin/gh/janeyx99/302/base 2025-09-07T07:36:00.0873867Z * [new branch] gh/janeyx99/302/head -> origin/gh/janeyx99/302/head 2025-09-07T07:36:00.0874536Z * [new branch] gh/janeyx99/303/base -> origin/gh/janeyx99/303/base 2025-09-07T07:36:00.0874935Z * [new branch] gh/janeyx99/303/head -> origin/gh/janeyx99/303/head 2025-09-07T07:36:00.0875839Z * [new branch] gh/janeyx99/88/base -> origin/gh/janeyx99/88/base 2025-09-07T07:36:00.0876458Z * [new branch] gh/janeyx99/88/head -> origin/gh/janeyx99/88/head 2025-09-07T07:36:00.0876961Z * [new branch] gh/janeyx99/88/orig -> origin/gh/janeyx99/88/orig 2025-09-07T07:36:00.0878045Z * [new branch] gh/jansel/360/base -> origin/gh/jansel/360/base 2025-09-07T07:36:00.0878609Z * [new branch] gh/jansel/360/head -> origin/gh/jansel/360/head 2025-09-07T07:36:00.0879291Z * [new branch] gh/jansel/451/base -> origin/gh/jansel/451/base 2025-09-07T07:36:00.0880217Z * [new branch] gh/jansel/451/head -> origin/gh/jansel/451/head 2025-09-07T07:36:00.0880826Z * [new branch] gh/jansel/451/orig -> origin/gh/jansel/451/orig 2025-09-07T07:36:00.0881481Z * [new branch] gh/jansel/462/base -> origin/gh/jansel/462/base 2025-09-07T07:36:00.0882100Z * [new branch] gh/jansel/462/head -> origin/gh/jansel/462/head 2025-09-07T07:36:00.0882453Z * [new branch] gh/jansel/462/orig -> origin/gh/jansel/462/orig 2025-09-07T07:36:00.0883537Z * [new branch] gh/jansel/531/base -> origin/gh/jansel/531/base 2025-09-07T07:36:00.0884186Z * [new branch] gh/jansel/531/head -> origin/gh/jansel/531/head 2025-09-07T07:36:00.0884610Z * [new branch] gh/jansel/531/orig -> origin/gh/jansel/531/orig 2025-09-07T07:36:00.0885761Z * [new branch] gh/jbschlosser/208/head -> origin/gh/jbschlosser/208/head 2025-09-07T07:36:00.0886510Z * [new branch] gh/jbschlosser/247/base -> origin/gh/jbschlosser/247/base 2025-09-07T07:36:00.0886963Z * [new branch] gh/jbschlosser/247/head -> origin/gh/jbschlosser/247/head 2025-09-07T07:36:00.0887587Z * [new branch] gh/jbschlosser/247/orig -> origin/gh/jbschlosser/247/orig 2025-09-07T07:36:00.0888367Z * [new branch] gh/jbschlosser/248/base -> origin/gh/jbschlosser/248/base 2025-09-07T07:36:00.0888990Z * [new branch] gh/jbschlosser/248/head -> origin/gh/jbschlosser/248/head 2025-09-07T07:36:00.0889358Z * [new branch] gh/jbschlosser/248/orig -> origin/gh/jbschlosser/248/orig 2025-09-07T07:36:00.0890235Z * [new branch] gh/jbschlosser/250/base -> origin/gh/jbschlosser/250/base 2025-09-07T07:36:00.0890670Z * [new branch] gh/jbschlosser/250/head -> origin/gh/jbschlosser/250/head 2025-09-07T07:36:00.0891576Z * [new branch] gh/jbschlosser/250/orig -> origin/gh/jbschlosser/250/orig 2025-09-07T07:36:00.0892486Z * [new branch] gh/jiayisunx/59/base -> origin/gh/jiayisunx/59/base 2025-09-07T07:36:00.0893125Z * [new branch] gh/jiayisunx/59/head -> origin/gh/jiayisunx/59/head 2025-09-07T07:36:00.0893668Z * [new branch] gh/jiayisunx/59/orig -> origin/gh/jiayisunx/59/orig 2025-09-07T07:36:00.0894432Z * [new branch] gh/jiayisunx/61/base -> origin/gh/jiayisunx/61/base 2025-09-07T07:36:00.0894867Z * [new branch] gh/jiayisunx/61/head -> origin/gh/jiayisunx/61/head 2025-09-07T07:36:00.0895509Z * [new branch] gh/jiayisunx/61/orig -> origin/gh/jiayisunx/61/orig 2025-09-07T07:36:00.0896257Z * [new branch] gh/jiayisunx/64/base -> origin/gh/jiayisunx/64/base 2025-09-07T07:36:00.0896673Z * [new branch] gh/jiayisunx/64/head -> origin/gh/jiayisunx/64/head 2025-09-07T07:36:00.0897305Z * [new branch] gh/jiayisunx/64/orig -> origin/gh/jiayisunx/64/orig 2025-09-07T07:36:00.0898015Z * [new branch] gh/jiayisunx/65/base -> origin/gh/jiayisunx/65/base 2025-09-07T07:36:00.0898567Z * [new branch] gh/jiayisunx/65/head -> origin/gh/jiayisunx/65/head 2025-09-07T07:36:00.0898998Z * [new branch] gh/jiayisunx/65/orig -> origin/gh/jiayisunx/65/orig 2025-09-07T07:36:00.0899858Z * [new branch] gh/jiayisunx/66/base -> origin/gh/jiayisunx/66/base 2025-09-07T07:36:00.0900453Z * [new branch] gh/jiayisunx/66/head -> origin/gh/jiayisunx/66/head 2025-09-07T07:36:00.0900995Z * [new branch] gh/jiayisunx/66/orig -> origin/gh/jiayisunx/66/orig 2025-09-07T07:36:00.0901672Z * [new branch] gh/jiayisunx/67/base -> origin/gh/jiayisunx/67/base 2025-09-07T07:36:00.0902211Z * [new branch] gh/jiayisunx/67/head -> origin/gh/jiayisunx/67/head 2025-09-07T07:36:00.0902645Z * [new branch] gh/jiayisunx/67/orig -> origin/gh/jiayisunx/67/orig 2025-09-07T07:36:00.0903562Z * [new branch] gh/jiayisunx/68/base -> origin/gh/jiayisunx/68/base 2025-09-07T07:36:00.0903934Z * [new branch] gh/jiayisunx/68/head -> origin/gh/jiayisunx/68/head 2025-09-07T07:36:00.0904512Z * [new branch] gh/jiayisunx/68/orig -> origin/gh/jiayisunx/68/orig 2025-09-07T07:36:00.0905274Z * [new branch] gh/jiayisunx/69/base -> origin/gh/jiayisunx/69/base 2025-09-07T07:36:00.0905763Z * [new branch] gh/jiayisunx/69/head -> origin/gh/jiayisunx/69/head 2025-09-07T07:36:00.0906363Z * [new branch] gh/jiayisunx/69/orig -> origin/gh/jiayisunx/69/orig 2025-09-07T07:36:00.0907081Z * [new branch] gh/jiayisunx/70/base -> origin/gh/jiayisunx/70/base 2025-09-07T07:36:00.0907713Z * [new branch] gh/jiayisunx/70/head -> origin/gh/jiayisunx/70/head 2025-09-07T07:36:00.0908146Z * [new branch] gh/jiayisunx/70/orig -> origin/gh/jiayisunx/70/orig 2025-09-07T07:36:00.0908940Z * [new branch] gh/jiayisunx/71/base -> origin/gh/jiayisunx/71/base 2025-09-07T07:36:00.0909780Z * [new branch] gh/jiayisunx/71/head -> origin/gh/jiayisunx/71/head 2025-09-07T07:36:00.0910225Z * [new branch] gh/jiayisunx/71/orig -> origin/gh/jiayisunx/71/orig 2025-09-07T07:36:00.0911017Z * [new branch] gh/jiayisunx/72/base -> origin/gh/jiayisunx/72/base 2025-09-07T07:36:00.0911561Z * [new branch] gh/jiayisunx/72/head -> origin/gh/jiayisunx/72/head 2025-09-07T07:36:00.0912116Z * [new branch] gh/jiayisunx/72/orig -> origin/gh/jiayisunx/72/orig 2025-09-07T07:36:00.0913346Z * [new branch] gh/jiayisunx/73/base -> origin/gh/jiayisunx/73/base 2025-09-07T07:36:00.0913890Z * [new branch] gh/jiayisunx/73/head -> origin/gh/jiayisunx/73/head 2025-09-07T07:36:00.0914471Z * [new branch] gh/jiayisunx/73/orig -> origin/gh/jiayisunx/73/orig 2025-09-07T07:36:00.0915525Z * [new branch] gh/jiayisunx/74/base -> origin/gh/jiayisunx/74/base 2025-09-07T07:36:00.0916109Z * [new branch] gh/jiayisunx/74/head -> origin/gh/jiayisunx/74/head 2025-09-07T07:36:00.0916657Z * [new branch] gh/jiayisunx/74/orig -> origin/gh/jiayisunx/74/orig 2025-09-07T07:36:00.0917339Z * [new branch] gh/jiayisunx/75/base -> origin/gh/jiayisunx/75/base 2025-09-07T07:36:00.0917813Z * [new branch] gh/jiayisunx/75/head -> origin/gh/jiayisunx/75/head 2025-09-07T07:36:00.0918440Z * [new branch] gh/jiayisunx/75/orig -> origin/gh/jiayisunx/75/orig 2025-09-07T07:36:00.0919035Z * [new branch] gh/jiayisunx/76/base -> origin/gh/jiayisunx/76/base 2025-09-07T07:36:00.0919401Z * [new branch] gh/jiayisunx/76/head -> origin/gh/jiayisunx/76/head 2025-09-07T07:36:00.0920018Z * [new branch] gh/jiayisunx/76/orig -> origin/gh/jiayisunx/76/orig 2025-09-07T07:36:00.0920918Z * [new branch] gh/jjwu@meta.com/1/base -> origin/gh/jjwu@meta.com/1/base 2025-09-07T07:36:00.0921309Z * [new branch] gh/jjwu@meta.com/1/head -> origin/gh/jjwu@meta.com/1/head 2025-09-07T07:36:00.0922436Z * [new branch] gh/justinchuby/111/base -> origin/gh/justinchuby/111/base 2025-09-07T07:36:00.0923057Z * [new branch] gh/justinchuby/111/head -> origin/gh/justinchuby/111/head 2025-09-07T07:36:00.0923662Z * [new branch] gh/justinchuby/111/orig -> origin/gh/justinchuby/111/orig 2025-09-07T07:36:00.0924373Z * [new branch] gh/justinchuby/112/base -> origin/gh/justinchuby/112/base 2025-09-07T07:36:00.0924764Z * [new branch] gh/justinchuby/112/head -> origin/gh/justinchuby/112/head 2025-09-07T07:36:00.0925396Z * [new branch] gh/justinchuby/112/orig -> origin/gh/justinchuby/112/orig 2025-09-07T07:36:00.0926106Z * [new branch] gh/justinchuby/113/base -> origin/gh/justinchuby/113/base 2025-09-07T07:36:00.0926665Z * [new branch] gh/justinchuby/113/head -> origin/gh/justinchuby/113/head 2025-09-07T07:36:00.0927312Z * [new branch] gh/justinchuby/113/orig -> origin/gh/justinchuby/113/orig 2025-09-07T07:36:00.0927986Z * [new branch] gh/justinchuby/114/base -> origin/gh/justinchuby/114/base 2025-09-07T07:36:00.0928452Z * [new branch] gh/justinchuby/114/head -> origin/gh/justinchuby/114/head 2025-09-07T07:36:00.0929076Z * [new branch] gh/justinchuby/114/orig -> origin/gh/justinchuby/114/orig 2025-09-07T07:36:00.0929824Z * [new branch] gh/justinchuby/115/base -> origin/gh/justinchuby/115/base 2025-09-07T07:36:00.0930268Z * [new branch] gh/justinchuby/115/head -> origin/gh/justinchuby/115/head 2025-09-07T07:36:00.0930821Z * [new branch] gh/justinchuby/115/orig -> origin/gh/justinchuby/115/orig 2025-09-07T07:36:00.0931771Z * [new branch] gh/karthickai/1/base -> origin/gh/karthickai/1/base 2025-09-07T07:36:00.0932222Z * [new branch] gh/karthickai/1/head -> origin/gh/karthickai/1/head 2025-09-07T07:36:00.0932998Z * [new branch] gh/karthickai/1/orig -> origin/gh/karthickai/1/orig 2025-09-07T07:36:00.0933674Z * [new branch] gh/karthickai/2/base -> origin/gh/karthickai/2/base 2025-09-07T07:36:00.0934030Z * [new branch] gh/karthickai/2/head -> origin/gh/karthickai/2/head 2025-09-07T07:36:00.0934651Z * [new branch] gh/karthickai/2/orig -> origin/gh/karthickai/2/orig 2025-09-07T07:36:00.0935569Z * [new branch] gh/kurtamohler/32/base -> origin/gh/kurtamohler/32/base 2025-09-07T07:36:00.0936171Z * [new branch] gh/kurtamohler/32/head -> origin/gh/kurtamohler/32/head 2025-09-07T07:36:00.0936624Z * [new branch] gh/kurtamohler/32/orig -> origin/gh/kurtamohler/32/orig 2025-09-07T07:36:00.0937402Z * [new branch] gh/kurtamohler/33/base -> origin/gh/kurtamohler/33/base 2025-09-07T07:36:00.0937842Z * [new branch] gh/kurtamohler/33/head -> origin/gh/kurtamohler/33/head 2025-09-07T07:36:00.0938442Z * [new branch] gh/kurtamohler/33/orig -> origin/gh/kurtamohler/33/orig 2025-09-07T07:36:00.0939490Z * [new branch] gh/kurtamohler/34/base -> origin/gh/kurtamohler/34/base 2025-09-07T07:36:00.0939989Z * [new branch] gh/kurtamohler/34/head -> origin/gh/kurtamohler/34/head 2025-09-07T07:36:00.0940533Z * [new branch] gh/kurtamohler/34/orig -> origin/gh/kurtamohler/34/orig 2025-09-07T07:36:00.0941361Z * [new branch] gh/kurtamohler/41/base -> origin/gh/kurtamohler/41/base 2025-09-07T07:36:00.0941800Z * [new branch] gh/kurtamohler/41/head -> origin/gh/kurtamohler/41/head 2025-09-07T07:36:00.0942387Z * [new branch] gh/kurtamohler/41/orig -> origin/gh/kurtamohler/41/orig 2025-09-07T07:36:00.0943099Z * [new branch] gh/kurtamohler/46/base -> origin/gh/kurtamohler/46/base 2025-09-07T07:36:00.0943532Z * [new branch] gh/kurtamohler/46/head -> origin/gh/kurtamohler/46/head 2025-09-07T07:36:00.0944199Z * [new branch] gh/kurtamohler/46/orig -> origin/gh/kurtamohler/46/orig 2025-09-07T07:36:00.0944952Z * [new branch] gh/kurtamohler/47/base -> origin/gh/kurtamohler/47/base 2025-09-07T07:36:00.0945514Z * [new branch] gh/kurtamohler/47/head -> origin/gh/kurtamohler/47/head 2025-09-07T07:36:00.0946064Z * [new branch] gh/kurtamohler/47/orig -> origin/gh/kurtamohler/47/orig 2025-09-07T07:36:00.0946768Z * [new branch] gh/kurtamohler/48/base -> origin/gh/kurtamohler/48/base 2025-09-07T07:36:00.0947372Z * [new branch] gh/kurtamohler/48/head -> origin/gh/kurtamohler/48/head 2025-09-07T07:36:00.0947750Z * [new branch] gh/kurtamohler/48/orig -> origin/gh/kurtamohler/48/orig 2025-09-07T07:36:00.0948560Z * [new branch] gh/kurtamohler/49/base -> origin/gh/kurtamohler/49/base 2025-09-07T07:36:00.0948997Z * [new branch] gh/kurtamohler/49/head -> origin/gh/kurtamohler/49/head 2025-09-07T07:36:00.0949538Z * [new branch] gh/kurtamohler/49/orig -> origin/gh/kurtamohler/49/orig 2025-09-07T07:36:00.0950333Z * [new branch] gh/kurtamohler/50/base -> origin/gh/kurtamohler/50/base 2025-09-07T07:36:00.0950888Z * [new branch] gh/kurtamohler/50/head -> origin/gh/kurtamohler/50/head 2025-09-07T07:36:00.0951504Z * [new branch] gh/kurtamohler/50/orig -> origin/gh/kurtamohler/50/orig 2025-09-07T07:36:00.0952576Z * [new branch] gh/kwen2501/130/base -> origin/gh/kwen2501/130/base 2025-09-07T07:36:00.0953232Z * [new branch] gh/kwen2501/130/head -> origin/gh/kwen2501/130/head 2025-09-07T07:36:00.0953782Z * [new branch] gh/kwen2501/130/orig -> origin/gh/kwen2501/130/orig 2025-09-07T07:36:00.0954558Z * [new branch] gh/kwen2501/15/base -> origin/gh/kwen2501/15/base 2025-09-07T07:36:00.0955088Z * [new branch] gh/kwen2501/15/head -> origin/gh/kwen2501/15/head 2025-09-07T07:36:00.0955838Z * [new branch] gh/kwen2501/156/base -> origin/gh/kwen2501/156/base 2025-09-07T07:36:00.0956282Z * [new branch] gh/kwen2501/156/head -> origin/gh/kwen2501/156/head 2025-09-07T07:36:00.0956884Z * [new branch] gh/kwen2501/156/orig -> origin/gh/kwen2501/156/orig 2025-09-07T07:36:00.0957639Z * [new branch] gh/kwen2501/170/base -> origin/gh/kwen2501/170/base 2025-09-07T07:36:00.0958064Z * [new branch] gh/kwen2501/170/head -> origin/gh/kwen2501/170/head 2025-09-07T07:36:00.0959023Z * [new branch] gh/kwen2501/186/base -> origin/gh/kwen2501/186/base 2025-09-07T07:36:00.0959458Z * [new branch] gh/kwen2501/186/head -> origin/gh/kwen2501/186/head 2025-09-07T07:36:00.0960203Z * [new branch] gh/kwen2501/186/orig -> origin/gh/kwen2501/186/orig 2025-09-07T07:36:00.0960880Z * [new branch] gh/kwen2501/187/base -> origin/gh/kwen2501/187/base 2025-09-07T07:36:00.0961530Z * [new branch] gh/kwen2501/187/head -> origin/gh/kwen2501/187/head 2025-09-07T07:36:00.0961969Z * [new branch] gh/kwen2501/187/orig -> origin/gh/kwen2501/187/orig 2025-09-07T07:36:00.0962816Z * [new branch] gh/kwen2501/188/base -> origin/gh/kwen2501/188/base 2025-09-07T07:36:00.0963357Z * [new branch] gh/kwen2501/188/head -> origin/gh/kwen2501/188/head 2025-09-07T07:36:00.0963797Z * [new branch] gh/kwen2501/188/orig -> origin/gh/kwen2501/188/orig 2025-09-07T07:36:00.0964625Z * [new branch] gh/kwen2501/194/base -> origin/gh/kwen2501/194/base 2025-09-07T07:36:00.0965062Z * [new branch] gh/kwen2501/194/head -> origin/gh/kwen2501/194/head 2025-09-07T07:36:00.0965660Z * [new branch] gh/kwen2501/194/orig -> origin/gh/kwen2501/194/orig 2025-09-07T07:36:00.0966449Z * [new branch] gh/kwen2501/199/base -> origin/gh/kwen2501/199/base 2025-09-07T07:36:00.0967051Z * [new branch] gh/kwen2501/199/head -> origin/gh/kwen2501/199/head 2025-09-07T07:36:00.0967422Z * [new branch] gh/kwen2501/199/orig -> origin/gh/kwen2501/199/orig 2025-09-07T07:36:00.0968137Z * [new branch] gh/kwen2501/200/base -> origin/gh/kwen2501/200/base 2025-09-07T07:36:00.0968886Z * [new branch] gh/kwen2501/200/head -> origin/gh/kwen2501/200/head 2025-09-07T07:36:00.0969327Z * [new branch] gh/kwen2501/200/orig -> origin/gh/kwen2501/200/orig 2025-09-07T07:36:00.0970092Z * [new branch] gh/kwen2501/201/base -> origin/gh/kwen2501/201/base 2025-09-07T07:36:00.0970541Z * [new branch] gh/kwen2501/201/head -> origin/gh/kwen2501/201/head 2025-09-07T07:36:00.0971135Z * [new branch] gh/kwen2501/201/orig -> origin/gh/kwen2501/201/orig 2025-09-07T07:36:00.0971836Z * [new branch] gh/kwen2501/203/base -> origin/gh/kwen2501/203/base 2025-09-07T07:36:00.0972471Z * [new branch] gh/kwen2501/203/head -> origin/gh/kwen2501/203/head 2025-09-07T07:36:00.0972802Z * [new branch] gh/kwen2501/203/orig -> origin/gh/kwen2501/203/orig 2025-09-07T07:36:00.0973672Z * [new branch] gh/kwen2501/204/base -> origin/gh/kwen2501/204/base 2025-09-07T07:36:00.0974086Z * [new branch] gh/kwen2501/204/head -> origin/gh/kwen2501/204/head 2025-09-07T07:36:00.0974689Z * [new branch] gh/kwen2501/204/orig -> origin/gh/kwen2501/204/orig 2025-09-07T07:36:00.0975405Z * [new branch] gh/kwen2501/205/base -> origin/gh/kwen2501/205/base 2025-09-07T07:36:00.0975846Z * [new branch] gh/kwen2501/205/head -> origin/gh/kwen2501/205/head 2025-09-07T07:36:00.0976437Z * [new branch] gh/kwen2501/205/orig -> origin/gh/kwen2501/205/orig 2025-09-07T07:36:00.0977132Z * [new branch] gh/kwen2501/206/base -> origin/gh/kwen2501/206/base 2025-09-07T07:36:00.0977770Z * [new branch] gh/kwen2501/206/head -> origin/gh/kwen2501/206/head 2025-09-07T07:36:00.0978202Z * [new branch] gh/kwen2501/206/orig -> origin/gh/kwen2501/206/orig 2025-09-07T07:36:00.0979078Z * [new branch] gh/kwen2501/207/base -> origin/gh/kwen2501/207/base 2025-09-07T07:36:00.0979438Z * [new branch] gh/kwen2501/207/head -> origin/gh/kwen2501/207/head 2025-09-07T07:36:00.0980043Z * [new branch] gh/kwen2501/207/orig -> origin/gh/kwen2501/207/orig 2025-09-07T07:36:00.0980813Z * [new branch] gh/kwen2501/208/base -> origin/gh/kwen2501/208/base 2025-09-07T07:36:00.0981252Z * [new branch] gh/kwen2501/208/head -> origin/gh/kwen2501/208/head 2025-09-07T07:36:00.0981840Z * [new branch] gh/kwen2501/208/orig -> origin/gh/kwen2501/208/orig 2025-09-07T07:36:00.0982897Z * [new branch] gh/kwen2501/209/base -> origin/gh/kwen2501/209/base 2025-09-07T07:36:00.0983286Z * [new branch] gh/kwen2501/209/head -> origin/gh/kwen2501/209/head 2025-09-07T07:36:00.0983877Z * [new branch] gh/kwen2501/209/orig -> origin/gh/kwen2501/209/orig 2025-09-07T07:36:00.0984658Z * [new branch] gh/kwen2501/210/base -> origin/gh/kwen2501/210/base 2025-09-07T07:36:00.0985095Z * [new branch] gh/kwen2501/210/head -> origin/gh/kwen2501/210/head 2025-09-07T07:36:00.0985720Z * [new branch] gh/kwen2501/210/orig -> origin/gh/kwen2501/210/orig 2025-09-07T07:36:00.0986472Z * [new branch] gh/kwen2501/211/base -> origin/gh/kwen2501/211/base 2025-09-07T07:36:00.0987966Z * [new branch] gh/kwen2501/211/head -> origin/gh/kwen2501/211/head 2025-09-07T07:36:00.0988165Z * [new branch] gh/kwen2501/212/base -> origin/gh/kwen2501/212/base 2025-09-07T07:36:00.0988413Z * [new branch] gh/kwen2501/212/head -> origin/gh/kwen2501/212/head 2025-09-07T07:36:00.0988854Z * [new branch] gh/kwen2501/212/orig -> origin/gh/kwen2501/212/orig 2025-09-07T07:36:00.0989661Z * [new branch] gh/kwen2501/213/base -> origin/gh/kwen2501/213/base 2025-09-07T07:36:00.0990161Z * [new branch] gh/kwen2501/213/head -> origin/gh/kwen2501/213/head 2025-09-07T07:36:00.0990732Z * [new branch] gh/kwen2501/213/orig -> origin/gh/kwen2501/213/orig 2025-09-07T07:36:00.0991499Z * [new branch] gh/kwen2501/214/base -> origin/gh/kwen2501/214/base 2025-09-07T07:36:00.0992043Z * [new branch] gh/kwen2501/214/head -> origin/gh/kwen2501/214/head 2025-09-07T07:36:00.0992587Z * [new branch] gh/kwen2501/214/orig -> origin/gh/kwen2501/214/orig 2025-09-07T07:36:00.0993376Z * [new branch] gh/kwen2501/215/base -> origin/gh/kwen2501/215/base 2025-09-07T07:36:00.0993758Z * [new branch] gh/kwen2501/215/head -> origin/gh/kwen2501/215/head 2025-09-07T07:36:00.0994424Z * [new branch] gh/kwen2501/215/orig -> origin/gh/kwen2501/215/orig 2025-09-07T07:36:00.0995079Z * [new branch] gh/kwen2501/216/base -> origin/gh/kwen2501/216/base 2025-09-07T07:36:00.0995687Z * [new branch] gh/kwen2501/216/head -> origin/gh/kwen2501/216/head 2025-09-07T07:36:00.0996247Z * [new branch] gh/kwen2501/216/orig -> origin/gh/kwen2501/216/orig 2025-09-07T07:36:00.0996973Z * [new branch] gh/kwen2501/217/base -> origin/gh/kwen2501/217/base 2025-09-07T07:36:00.0997406Z * [new branch] gh/kwen2501/217/head -> origin/gh/kwen2501/217/head 2025-09-07T07:36:00.0998040Z * [new branch] gh/kwen2501/217/orig -> origin/gh/kwen2501/217/orig 2025-09-07T07:36:00.0998699Z * [new branch] gh/kwen2501/218/base -> origin/gh/kwen2501/218/base 2025-09-07T07:36:00.0999152Z * [new branch] gh/kwen2501/218/head -> origin/gh/kwen2501/218/head 2025-09-07T07:36:00.0999752Z * [new branch] gh/kwen2501/218/orig -> origin/gh/kwen2501/218/orig 2025-09-07T07:36:00.1000450Z * [new branch] gh/kwen2501/219/base -> origin/gh/kwen2501/219/base 2025-09-07T07:36:00.1000902Z * [new branch] gh/kwen2501/219/head -> origin/gh/kwen2501/219/head 2025-09-07T07:36:00.1001572Z * [new branch] gh/kwen2501/219/orig -> origin/gh/kwen2501/219/orig 2025-09-07T07:36:00.1002292Z * [new branch] gh/kwen2501/220/base -> origin/gh/kwen2501/220/base 2025-09-07T07:36:00.1003179Z * [new branch] gh/kwen2501/220/head -> origin/gh/kwen2501/220/head 2025-09-07T07:36:00.1003582Z * [new branch] gh/kwen2501/220/orig -> origin/gh/kwen2501/220/orig 2025-09-07T07:36:00.1004427Z * [new branch] gh/kwen2501/221/base -> origin/gh/kwen2501/221/base 2025-09-07T07:36:00.1004874Z * [new branch] gh/kwen2501/221/head -> origin/gh/kwen2501/221/head 2025-09-07T07:36:00.1005566Z * [new branch] gh/kwen2501/221/orig -> origin/gh/kwen2501/221/orig 2025-09-07T07:36:00.1006302Z * [new branch] gh/kwen2501/222/base -> origin/gh/kwen2501/222/base 2025-09-07T07:36:00.1006744Z * [new branch] gh/kwen2501/222/head -> origin/gh/kwen2501/222/head 2025-09-07T07:36:00.1007328Z * [new branch] gh/kwen2501/222/orig -> origin/gh/kwen2501/222/orig 2025-09-07T07:36:00.1008002Z * [new branch] gh/kwen2501/223/base -> origin/gh/kwen2501/223/base 2025-09-07T07:36:00.1008433Z * [new branch] gh/kwen2501/223/head -> origin/gh/kwen2501/223/head 2025-09-07T07:36:00.1009030Z * [new branch] gh/kwen2501/223/orig -> origin/gh/kwen2501/223/orig 2025-09-07T07:36:00.1009804Z * [new branch] gh/kwen2501/224/base -> origin/gh/kwen2501/224/base 2025-09-07T07:36:00.1010363Z * [new branch] gh/kwen2501/224/head -> origin/gh/kwen2501/224/head 2025-09-07T07:36:00.1010769Z * [new branch] gh/kwen2501/224/orig -> origin/gh/kwen2501/224/orig 2025-09-07T07:36:00.1011569Z * [new branch] gh/kwen2501/225/base -> origin/gh/kwen2501/225/base 2025-09-07T07:36:00.1012004Z * [new branch] gh/kwen2501/225/head -> origin/gh/kwen2501/225/head 2025-09-07T07:36:00.1012588Z * [new branch] gh/kwen2501/225/orig -> origin/gh/kwen2501/225/orig 2025-09-07T07:36:00.1013253Z * [new branch] gh/kwen2501/226/base -> origin/gh/kwen2501/226/base 2025-09-07T07:36:00.1013895Z * [new branch] gh/kwen2501/226/head -> origin/gh/kwen2501/226/head 2025-09-07T07:36:00.1014368Z * [new branch] gh/kwen2501/226/orig -> origin/gh/kwen2501/226/orig 2025-09-07T07:36:00.1015134Z * [new branch] gh/kwen2501/227/base -> origin/gh/kwen2501/227/base 2025-09-07T07:36:00.1015548Z * [new branch] gh/kwen2501/227/head -> origin/gh/kwen2501/227/head 2025-09-07T07:36:00.1016131Z * [new branch] gh/kwen2501/227/orig -> origin/gh/kwen2501/227/orig 2025-09-07T07:36:00.1016940Z * [new branch] gh/kwen2501/228/base -> origin/gh/kwen2501/228/base 2025-09-07T07:36:00.1017370Z * [new branch] gh/kwen2501/228/head -> origin/gh/kwen2501/228/head 2025-09-07T07:36:00.1017973Z * [new branch] gh/kwen2501/228/orig -> origin/gh/kwen2501/228/orig 2025-09-07T07:36:00.1018672Z * [new branch] gh/kwen2501/229/base -> origin/gh/kwen2501/229/base 2025-09-07T07:36:00.1019245Z * [new branch] gh/kwen2501/229/head -> origin/gh/kwen2501/229/head 2025-09-07T07:36:00.1019630Z * [new branch] gh/kwen2501/229/orig -> origin/gh/kwen2501/229/orig 2025-09-07T07:36:00.1020474Z * [new branch] gh/kwen2501/230/base -> origin/gh/kwen2501/230/base 2025-09-07T07:36:00.1020851Z * [new branch] gh/kwen2501/230/head -> origin/gh/kwen2501/230/head 2025-09-07T07:36:00.1021512Z * [new branch] gh/kwen2501/230/orig -> origin/gh/kwen2501/230/orig 2025-09-07T07:36:00.1022229Z * [new branch] gh/kwen2501/231/base -> origin/gh/kwen2501/231/base 2025-09-07T07:36:00.1022944Z * [new branch] gh/kwen2501/231/head -> origin/gh/kwen2501/231/head 2025-09-07T07:36:00.1023340Z * [new branch] gh/kwen2501/231/orig -> origin/gh/kwen2501/231/orig 2025-09-07T07:36:00.1024195Z * [new branch] gh/kwen2501/232/base -> origin/gh/kwen2501/232/base 2025-09-07T07:36:00.1024746Z * [new branch] gh/kwen2501/232/head -> origin/gh/kwen2501/232/head 2025-09-07T07:36:00.1025184Z * [new branch] gh/kwen2501/232/orig -> origin/gh/kwen2501/232/orig 2025-09-07T07:36:00.1026391Z * [new branch] gh/laithsakka/156/base -> origin/gh/laithsakka/156/base 2025-09-07T07:36:00.1026783Z * [new branch] gh/laithsakka/156/head -> origin/gh/laithsakka/156/head 2025-09-07T07:36:00.1027395Z * [new branch] gh/laithsakka/156/orig -> origin/gh/laithsakka/156/orig 2025-09-07T07:36:00.1028229Z * [new branch] gh/laithsakka/160/base -> origin/gh/laithsakka/160/base 2025-09-07T07:36:00.1028616Z * [new branch] gh/laithsakka/160/head -> origin/gh/laithsakka/160/head 2025-09-07T07:36:00.1029234Z * [new branch] gh/laithsakka/160/orig -> origin/gh/laithsakka/160/orig 2025-09-07T07:36:00.1029944Z * [new branch] gh/laithsakka/178/base -> origin/gh/laithsakka/178/base 2025-09-07T07:36:00.1030547Z * [new branch] gh/laithsakka/178/head -> origin/gh/laithsakka/178/head 2025-09-07T07:36:00.1030974Z * [new branch] gh/laithsakka/178/orig -> origin/gh/laithsakka/178/orig 2025-09-07T07:36:00.1031941Z * [new branch] gh/laithsakka/191/base -> origin/gh/laithsakka/191/base 2025-09-07T07:36:00.1032489Z * [new branch] gh/laithsakka/191/head -> origin/gh/laithsakka/191/head 2025-09-07T07:36:00.1032924Z * [new branch] gh/laithsakka/191/orig -> origin/gh/laithsakka/191/orig 2025-09-07T07:36:00.1033770Z * [new branch] gh/laithsakka/237/base -> origin/gh/laithsakka/237/base 2025-09-07T07:36:00.1034141Z * [new branch] gh/laithsakka/237/head -> origin/gh/laithsakka/237/head 2025-09-07T07:36:00.1034734Z * [new branch] gh/laithsakka/237/orig -> origin/gh/laithsakka/237/orig 2025-09-07T07:36:00.1035468Z * [new branch] gh/laithsakka/249/base -> origin/gh/laithsakka/249/base 2025-09-07T07:36:00.1036342Z * [new branch] gh/laithsakka/249/head -> origin/gh/laithsakka/249/head 2025-09-07T07:36:00.1036774Z * [new branch] gh/laithsakka/249/orig -> origin/gh/laithsakka/249/orig 2025-09-07T07:36:00.1037627Z * [new branch] gh/laithsakka/251/base -> origin/gh/laithsakka/251/base 2025-09-07T07:36:00.1038070Z * [new branch] gh/laithsakka/251/head -> origin/gh/laithsakka/251/head 2025-09-07T07:36:00.1038722Z * [new branch] gh/laithsakka/251/orig -> origin/gh/laithsakka/251/orig 2025-09-07T07:36:00.1039519Z * [new branch] gh/laithsakka/254/base -> origin/gh/laithsakka/254/base 2025-09-07T07:36:00.1039929Z * [new branch] gh/laithsakka/254/head -> origin/gh/laithsakka/254/head 2025-09-07T07:36:00.1040643Z * [new branch] gh/laithsakka/254/orig -> origin/gh/laithsakka/254/orig 2025-09-07T07:36:00.1041364Z * [new branch] gh/laithsakka/255/base -> origin/gh/laithsakka/255/base 2025-09-07T07:36:00.1041926Z * [new branch] gh/laithsakka/255/head -> origin/gh/laithsakka/255/head 2025-09-07T07:36:00.1042338Z * [new branch] gh/laithsakka/255/orig -> origin/gh/laithsakka/255/orig 2025-09-07T07:36:00.1043142Z * [new branch] gh/laithsakka/256/base -> origin/gh/laithsakka/256/base 2025-09-07T07:36:00.1043568Z * [new branch] gh/laithsakka/256/head -> origin/gh/laithsakka/256/head 2025-09-07T07:36:00.1044118Z * [new branch] gh/laithsakka/256/orig -> origin/gh/laithsakka/256/orig 2025-09-07T07:36:00.1044845Z * [new branch] gh/laithsakka/257/base -> origin/gh/laithsakka/257/base 2025-09-07T07:36:00.1045405Z * [new branch] gh/laithsakka/257/head -> origin/gh/laithsakka/257/head 2025-09-07T07:36:00.1045834Z * [new branch] gh/laithsakka/257/orig -> origin/gh/laithsakka/257/orig 2025-09-07T07:36:00.1046810Z * [new branch] gh/laithsakka/258/base -> origin/gh/laithsakka/258/base 2025-09-07T07:36:00.1047224Z * [new branch] gh/laithsakka/258/head -> origin/gh/laithsakka/258/head 2025-09-07T07:36:00.1047792Z * [new branch] gh/laithsakka/258/orig -> origin/gh/laithsakka/258/orig 2025-09-07T07:36:00.1048608Z * [new branch] gh/laithsakka/259/base -> origin/gh/laithsakka/259/base 2025-09-07T07:36:00.1048997Z * [new branch] gh/laithsakka/259/head -> origin/gh/laithsakka/259/head 2025-09-07T07:36:00.1049600Z * [new branch] gh/laithsakka/259/orig -> origin/gh/laithsakka/259/orig 2025-09-07T07:36:00.1050267Z * [new branch] gh/laithsakka/260/base -> origin/gh/laithsakka/260/base 2025-09-07T07:36:00.1050948Z * [new branch] gh/laithsakka/260/head -> origin/gh/laithsakka/260/head 2025-09-07T07:36:00.1051395Z * [new branch] gh/laithsakka/260/orig -> origin/gh/laithsakka/260/orig 2025-09-07T07:36:00.1052251Z * [new branch] gh/laithsakka/261/base -> origin/gh/laithsakka/261/base 2025-09-07T07:36:00.1052665Z * [new branch] gh/laithsakka/261/head -> origin/gh/laithsakka/261/head 2025-09-07T07:36:00.1053285Z * [new branch] gh/laithsakka/261/orig -> origin/gh/laithsakka/261/orig 2025-09-07T07:36:00.1054339Z * [new branch] gh/laithsakka/262/base -> origin/gh/laithsakka/262/base 2025-09-07T07:36:00.1055175Z * [new branch] gh/laithsakka/262/head -> origin/gh/laithsakka/262/head 2025-09-07T07:36:00.1055618Z * [new branch] gh/laithsakka/262/orig -> origin/gh/laithsakka/262/orig 2025-09-07T07:36:00.1056700Z * [new branch] gh/laithsakka/263/base -> origin/gh/laithsakka/263/base 2025-09-07T07:36:00.1057137Z * [new branch] gh/laithsakka/263/head -> origin/gh/laithsakka/263/head 2025-09-07T07:36:00.1057818Z * [new branch] gh/laithsakka/263/orig -> origin/gh/laithsakka/263/orig 2025-09-07T07:36:00.1058524Z * [new branch] gh/laithsakka/264/base -> origin/gh/laithsakka/264/base 2025-09-07T07:36:00.1058936Z * [new branch] gh/laithsakka/264/head -> origin/gh/laithsakka/264/head 2025-09-07T07:36:00.1059494Z * [new branch] gh/laithsakka/264/orig -> origin/gh/laithsakka/264/orig 2025-09-07T07:36:00.1060426Z * [new branch] gh/laithsakka/265/base -> origin/gh/laithsakka/265/base 2025-09-07T07:36:00.1060995Z * [new branch] gh/laithsakka/265/head -> origin/gh/laithsakka/265/head 2025-09-07T07:36:00.1061401Z * [new branch] gh/laithsakka/265/orig -> origin/gh/laithsakka/265/orig 2025-09-07T07:36:00.1062372Z * [new branch] gh/laithsakka/266/base -> origin/gh/laithsakka/266/base 2025-09-07T07:36:00.1062715Z * [new branch] gh/laithsakka/266/head -> origin/gh/laithsakka/266/head 2025-09-07T07:36:00.1063315Z * [new branch] gh/laithsakka/266/orig -> origin/gh/laithsakka/266/orig 2025-09-07T07:36:00.1064388Z * [new branch] gh/laithsakka/267/base -> origin/gh/laithsakka/267/base 2025-09-07T07:36:00.1064823Z * [new branch] gh/laithsakka/267/head -> origin/gh/laithsakka/267/head 2025-09-07T07:36:00.1065434Z * [new branch] gh/laithsakka/267/orig -> origin/gh/laithsakka/267/orig 2025-09-07T07:36:00.1066155Z * [new branch] gh/laithsakka/268/base -> origin/gh/laithsakka/268/base 2025-09-07T07:36:00.1066585Z * [new branch] gh/laithsakka/268/head -> origin/gh/laithsakka/268/head 2025-09-07T07:36:00.1067312Z * [new branch] gh/laithsakka/268/orig -> origin/gh/laithsakka/268/orig 2025-09-07T07:36:00.1068050Z * [new branch] gh/laithsakka/28/base -> origin/gh/laithsakka/28/base 2025-09-07T07:36:00.1068820Z * [new branch] gh/laithsakka/29/base -> origin/gh/laithsakka/29/base 2025-09-07T07:36:00.1069615Z * [new branch] gh/laithsakka/30/base -> origin/gh/laithsakka/30/base 2025-09-07T07:36:00.1070075Z * [new branch] gh/laithsakka/30/head -> origin/gh/laithsakka/30/head 2025-09-07T07:36:00.1070827Z * [new branch] gh/laithsakka/31/base -> origin/gh/laithsakka/31/base 2025-09-07T07:36:00.1071174Z * [new branch] gh/laithsakka/31/head -> origin/gh/laithsakka/31/head 2025-09-07T07:36:00.1071915Z * [new branch] gh/laithsakka/32/base -> origin/gh/laithsakka/32/base 2025-09-07T07:36:00.1072304Z * [new branch] gh/laithsakka/32/head -> origin/gh/laithsakka/32/head 2025-09-07T07:36:00.1074735Z * [new branch] gh/lucaskabela/1/base -> origin/gh/lucaskabela/1/base 2025-09-07T07:36:00.1075288Z * [new branch] gh/lucaskabela/1/head -> origin/gh/lucaskabela/1/head 2025-09-07T07:36:00.1076479Z * [new branch] gh/lucaskabela/10/base -> origin/gh/lucaskabela/10/base 2025-09-07T07:36:00.1077036Z * [new branch] gh/lucaskabela/10/head -> origin/gh/lucaskabela/10/head 2025-09-07T07:36:00.1077642Z * [new branch] gh/lucaskabela/10/orig -> origin/gh/lucaskabela/10/orig 2025-09-07T07:36:00.1078261Z * [new branch] gh/lucaskabela/11/base -> origin/gh/lucaskabela/11/base 2025-09-07T07:36:00.1078956Z * [new branch] gh/lucaskabela/11/head -> origin/gh/lucaskabela/11/head 2025-09-07T07:36:00.1079253Z * [new branch] gh/lucaskabela/11/orig -> origin/gh/lucaskabela/11/orig 2025-09-07T07:36:00.1079923Z * [new branch] gh/lucaskabela/12/base -> origin/gh/lucaskabela/12/base 2025-09-07T07:36:00.1080378Z * [new branch] gh/lucaskabela/12/head -> origin/gh/lucaskabela/12/head 2025-09-07T07:36:00.1081059Z * [new branch] gh/lucaskabela/12/orig -> origin/gh/lucaskabela/12/orig 2025-09-07T07:36:00.1081704Z * [new branch] gh/lucaskabela/13/base -> origin/gh/lucaskabela/13/base 2025-09-07T07:36:00.1082202Z * [new branch] gh/lucaskabela/13/head -> origin/gh/lucaskabela/13/head 2025-09-07T07:36:00.1083103Z * [new branch] gh/lucaskabela/13/orig -> origin/gh/lucaskabela/13/orig 2025-09-07T07:36:00.1083768Z * [new branch] gh/lucaskabela/14/base -> origin/gh/lucaskabela/14/base 2025-09-07T07:36:00.1084350Z * [new branch] gh/lucaskabela/14/head -> origin/gh/lucaskabela/14/head 2025-09-07T07:36:00.1084787Z * [new branch] gh/lucaskabela/14/orig -> origin/gh/lucaskabela/14/orig 2025-09-07T07:36:00.1085533Z * [new branch] gh/lucaskabela/15/base -> origin/gh/lucaskabela/15/base 2025-09-07T07:36:00.1086081Z * [new branch] gh/lucaskabela/15/head -> origin/gh/lucaskabela/15/head 2025-09-07T07:36:00.1086528Z * [new branch] gh/lucaskabela/15/orig -> origin/gh/lucaskabela/15/orig 2025-09-07T07:36:00.1087312Z * [new branch] gh/lucaskabela/16/base -> origin/gh/lucaskabela/16/base 2025-09-07T07:36:00.1087682Z * [new branch] gh/lucaskabela/16/head -> origin/gh/lucaskabela/16/head 2025-09-07T07:36:00.1088293Z * [new branch] gh/lucaskabela/16/orig -> origin/gh/lucaskabela/16/orig 2025-09-07T07:36:00.1088908Z * [new branch] gh/lucaskabela/17/base -> origin/gh/lucaskabela/17/base 2025-09-07T07:36:00.1089460Z * [new branch] gh/lucaskabela/17/head -> origin/gh/lucaskabela/17/head 2025-09-07T07:36:00.1090341Z * [new branch] gh/lucaskabela/17/orig -> origin/gh/lucaskabela/17/orig 2025-09-07T07:36:00.1091078Z * [new branch] gh/lucaskabela/2/base -> origin/gh/lucaskabela/2/base 2025-09-07T07:36:00.1091648Z * [new branch] gh/lucaskabela/2/head -> origin/gh/lucaskabela/2/head 2025-09-07T07:36:00.1092197Z * [new branch] gh/lucaskabela/2/orig -> origin/gh/lucaskabela/2/orig 2025-09-07T07:36:00.1093053Z * [new branch] gh/lucaskabela/3/base -> origin/gh/lucaskabela/3/base 2025-09-07T07:36:00.1093457Z * [new branch] gh/lucaskabela/3/head -> origin/gh/lucaskabela/3/head 2025-09-07T07:36:00.1094066Z * [new branch] gh/lucaskabela/3/orig -> origin/gh/lucaskabela/3/orig 2025-09-07T07:36:00.1094913Z * [new branch] gh/lucaskabela/4/base -> origin/gh/lucaskabela/4/base 2025-09-07T07:36:00.1095229Z * [new branch] gh/lucaskabela/4/head -> origin/gh/lucaskabela/4/head 2025-09-07T07:36:00.1095825Z * [new branch] gh/lucaskabela/4/orig -> origin/gh/lucaskabela/4/orig 2025-09-07T07:36:00.1096601Z * [new branch] gh/lucaskabela/5/base -> origin/gh/lucaskabela/5/base 2025-09-07T07:36:00.1097007Z * [new branch] gh/lucaskabela/5/head -> origin/gh/lucaskabela/5/head 2025-09-07T07:36:00.1097641Z * [new branch] gh/lucaskabela/5/orig -> origin/gh/lucaskabela/5/orig 2025-09-07T07:36:00.1098300Z * [new branch] gh/lucaskabela/6/base -> origin/gh/lucaskabela/6/base 2025-09-07T07:36:00.1098745Z * [new branch] gh/lucaskabela/6/head -> origin/gh/lucaskabela/6/head 2025-09-07T07:36:00.1099504Z * [new branch] gh/lucaskabela/6/orig -> origin/gh/lucaskabela/6/orig 2025-09-07T07:36:00.1100308Z * [new branch] gh/lucaskabela/7/base -> origin/gh/lucaskabela/7/base 2025-09-07T07:36:00.1100702Z * [new branch] gh/lucaskabela/7/head -> origin/gh/lucaskabela/7/head 2025-09-07T07:36:00.1101290Z * [new branch] gh/lucaskabela/7/orig -> origin/gh/lucaskabela/7/orig 2025-09-07T07:36:00.1101967Z * [new branch] gh/lucaskabela/8/base -> origin/gh/lucaskabela/8/base 2025-09-07T07:36:00.1102481Z * [new branch] gh/lucaskabela/8/head -> origin/gh/lucaskabela/8/head 2025-09-07T07:36:00.1103055Z * [new branch] gh/lucaskabela/8/orig -> origin/gh/lucaskabela/8/orig 2025-09-07T07:36:00.1103814Z * [new branch] gh/lucaskabela/9/base -> origin/gh/lucaskabela/9/base 2025-09-07T07:36:00.1104376Z * [new branch] gh/lucaskabela/9/head -> origin/gh/lucaskabela/9/head 2025-09-07T07:36:00.1104791Z * [new branch] gh/lucaskabela/9/orig -> origin/gh/lucaskabela/9/orig 2025-09-07T07:36:00.1105812Z * [new branch] gh/lw/3/base -> origin/gh/lw/3/base 2025-09-07T07:36:00.1106418Z * [new branch] gh/lw/3/head -> origin/gh/lw/3/head 2025-09-07T07:36:00.1106941Z * [new branch] gh/lw/3/orig -> origin/gh/lw/3/orig 2025-09-07T07:36:00.1107867Z * [new branch] gh/malfet/14/base -> origin/gh/malfet/14/base 2025-09-07T07:36:00.1108737Z * [new branch] gh/malfet/330/base -> origin/gh/malfet/330/base 2025-09-07T07:36:00.1109176Z * [new branch] gh/malfet/330/head -> origin/gh/malfet/330/head 2025-09-07T07:36:00.1109777Z * [new branch] gh/malfet/330/orig -> origin/gh/malfet/330/orig 2025-09-07T07:36:00.1110544Z * [new branch] gh/malfet/396/base -> origin/gh/malfet/396/base 2025-09-07T07:36:00.1111080Z * [new branch] gh/malfet/396/head -> origin/gh/malfet/396/head 2025-09-07T07:36:00.1111618Z * [new branch] gh/malfet/396/orig -> origin/gh/malfet/396/orig 2025-09-07T07:36:00.1112333Z * [new branch] gh/malfet/397/base -> origin/gh/malfet/397/base 2025-09-07T07:36:00.1112884Z * [new branch] gh/malfet/397/head -> origin/gh/malfet/397/head 2025-09-07T07:36:00.1113511Z * [new branch] gh/malfet/397/orig -> origin/gh/malfet/397/orig 2025-09-07T07:36:00.1114318Z * [new branch] gh/malfet/398/base -> origin/gh/malfet/398/base 2025-09-07T07:36:00.1114621Z * [new branch] gh/malfet/398/head -> origin/gh/malfet/398/head 2025-09-07T07:36:00.1115204Z * [new branch] gh/malfet/398/orig -> origin/gh/malfet/398/orig 2025-09-07T07:36:00.1115947Z * [new branch] gh/malfet/399/base -> origin/gh/malfet/399/base 2025-09-07T07:36:00.1116381Z * [new branch] gh/malfet/399/head -> origin/gh/malfet/399/head 2025-09-07T07:36:00.1116985Z * [new branch] gh/malfet/399/orig -> origin/gh/malfet/399/orig 2025-09-07T07:36:00.1117865Z * [new branch] gh/malfet/414/base -> origin/gh/malfet/414/base 2025-09-07T07:36:00.1119500Z * [new branch] gh/malfet/414/head -> origin/gh/malfet/414/head 2025-09-07T07:36:00.1119624Z * [new branch] gh/malfet/414/orig -> origin/gh/malfet/414/orig 2025-09-07T07:36:00.1119858Z * [new branch] gh/malfet/417/base -> origin/gh/malfet/417/base 2025-09-07T07:36:00.1121661Z * [new branch] gh/malfet/417/head -> origin/gh/malfet/417/head 2025-09-07T07:36:00.1121786Z * [new branch] gh/malfet/417/orig -> origin/gh/malfet/417/orig 2025-09-07T07:36:00.1121906Z * [new branch] gh/malfet/418/base -> origin/gh/malfet/418/base 2025-09-07T07:36:00.1122115Z * [new branch] gh/malfet/418/head -> origin/gh/malfet/418/head 2025-09-07T07:36:00.1122702Z * [new branch] gh/malfet/418/orig -> origin/gh/malfet/418/orig 2025-09-07T07:36:00.1123588Z * [new branch] gh/malfet/475/base -> origin/gh/malfet/475/base 2025-09-07T07:36:00.1123975Z * [new branch] gh/malfet/475/head -> origin/gh/malfet/475/head 2025-09-07T07:36:00.1124590Z * [new branch] gh/malfet/475/orig -> origin/gh/malfet/475/orig 2025-09-07T07:36:00.1125320Z * [new branch] gh/malfet/476/base -> origin/gh/malfet/476/base 2025-09-07T07:36:00.1125758Z * [new branch] gh/malfet/476/head -> origin/gh/malfet/476/head 2025-09-07T07:36:00.1126376Z * [new branch] gh/malfet/476/orig -> origin/gh/malfet/476/orig 2025-09-07T07:36:00.1127221Z * [new branch] gh/malfet/477/base -> origin/gh/malfet/477/base 2025-09-07T07:36:00.1127634Z * [new branch] gh/malfet/477/head -> origin/gh/malfet/477/head 2025-09-07T07:36:00.1128407Z * [new branch] gh/malfet/477/orig -> origin/gh/malfet/477/orig 2025-09-07T07:36:00.1129035Z * [new branch] gh/malfet/478/base -> origin/gh/malfet/478/base 2025-09-07T07:36:00.1129568Z * [new branch] gh/malfet/478/head -> origin/gh/malfet/478/head 2025-09-07T07:36:00.1130109Z * [new branch] gh/malfet/478/orig -> origin/gh/malfet/478/orig 2025-09-07T07:36:00.1130793Z * [new branch] gh/malfet/479/base -> origin/gh/malfet/479/base 2025-09-07T07:36:00.1131374Z * [new branch] gh/malfet/479/head -> origin/gh/malfet/479/head 2025-09-07T07:36:00.1131851Z * [new branch] gh/malfet/479/orig -> origin/gh/malfet/479/orig 2025-09-07T07:36:00.1132635Z * [new branch] gh/malfet/480/base -> origin/gh/malfet/480/base 2025-09-07T07:36:00.1133183Z * [new branch] gh/malfet/480/head -> origin/gh/malfet/480/head 2025-09-07T07:36:00.1133722Z * [new branch] gh/malfet/480/orig -> origin/gh/malfet/480/orig 2025-09-07T07:36:00.1134471Z * [new branch] gh/malfet/481/base -> origin/gh/malfet/481/base 2025-09-07T07:36:00.1135016Z * [new branch] gh/malfet/481/head -> origin/gh/malfet/481/head 2025-09-07T07:36:00.1135616Z * [new branch] gh/malfet/481/orig -> origin/gh/malfet/481/orig 2025-09-07T07:36:00.1136359Z * [new branch] gh/malfet/482/base -> origin/gh/malfet/482/base 2025-09-07T07:36:00.1136874Z * [new branch] gh/malfet/482/head -> origin/gh/malfet/482/head 2025-09-07T07:36:00.1137760Z * [new branch] gh/malfet/482/orig -> origin/gh/malfet/482/orig 2025-09-07T07:36:00.1138782Z * [new branch] gh/malfet/483/base -> origin/gh/malfet/483/base 2025-09-07T07:36:00.1139169Z * [new branch] gh/malfet/483/head -> origin/gh/malfet/483/head 2025-09-07T07:36:00.1139765Z * [new branch] gh/malfet/483/orig -> origin/gh/malfet/483/orig 2025-09-07T07:36:00.1140617Z * [new branch] gh/malfet/484/base -> origin/gh/malfet/484/base 2025-09-07T07:36:00.1140937Z * [new branch] gh/malfet/484/head -> origin/gh/malfet/484/head 2025-09-07T07:36:00.1141648Z * [new branch] gh/malfet/484/orig -> origin/gh/malfet/484/orig 2025-09-07T07:36:00.1142342Z * [new branch] gh/malfet/485/base -> origin/gh/malfet/485/base 2025-09-07T07:36:00.1142779Z * [new branch] gh/malfet/485/head -> origin/gh/malfet/485/head 2025-09-07T07:36:00.1143457Z * [new branch] gh/malfet/485/orig -> origin/gh/malfet/485/orig 2025-09-07T07:36:00.1144192Z * [new branch] gh/malfet/486/base -> origin/gh/malfet/486/base 2025-09-07T07:36:00.1144612Z * [new branch] gh/malfet/486/head -> origin/gh/malfet/486/head 2025-09-07T07:36:00.1145222Z * [new branch] gh/malfet/486/orig -> origin/gh/malfet/486/orig 2025-09-07T07:36:00.1146016Z * [new branch] gh/malfet/487/base -> origin/gh/malfet/487/base 2025-09-07T07:36:00.1146554Z * [new branch] gh/malfet/487/head -> origin/gh/malfet/487/head 2025-09-07T07:36:00.1147022Z * [new branch] gh/malfet/487/orig -> origin/gh/malfet/487/orig 2025-09-07T07:36:00.1147884Z * [new branch] gh/malfet/488/base -> origin/gh/malfet/488/base 2025-09-07T07:36:00.1148429Z * [new branch] gh/malfet/488/head -> origin/gh/malfet/488/head 2025-09-07T07:36:00.1149010Z * [new branch] gh/malfet/488/orig -> origin/gh/malfet/488/orig 2025-09-07T07:36:00.1149701Z * [new branch] gh/malfet/489/base -> origin/gh/malfet/489/base 2025-09-07T07:36:00.1150236Z * [new branch] gh/malfet/489/head -> origin/gh/malfet/489/head 2025-09-07T07:36:00.1151018Z * [new branch] gh/malfet/489/orig -> origin/gh/malfet/489/orig 2025-09-07T07:36:00.1151692Z * [new branch] gh/malfet/490/base -> origin/gh/malfet/490/base 2025-09-07T07:36:00.1152229Z * [new branch] gh/malfet/490/head -> origin/gh/malfet/490/head 2025-09-07T07:36:00.1152774Z * [new branch] gh/malfet/490/orig -> origin/gh/malfet/490/orig 2025-09-07T07:36:00.1153573Z * [new branch] gh/malfet/491/base -> origin/gh/malfet/491/base 2025-09-07T07:36:00.1154121Z * [new branch] gh/malfet/491/head -> origin/gh/malfet/491/head 2025-09-07T07:36:00.1154707Z * [new branch] gh/malfet/491/orig -> origin/gh/malfet/491/orig 2025-09-07T07:36:00.1155465Z * [new branch] gh/malfet/492/base -> origin/gh/malfet/492/base 2025-09-07T07:36:00.1155999Z * [new branch] gh/malfet/492/head -> origin/gh/malfet/492/head 2025-09-07T07:36:00.1156957Z * [new branch] gh/malfet/492/orig -> origin/gh/malfet/492/orig 2025-09-07T07:36:00.1157755Z * [new branch] gh/malfet/493/base -> origin/gh/malfet/493/base 2025-09-07T07:36:00.1158153Z * [new branch] gh/malfet/493/head -> origin/gh/malfet/493/head 2025-09-07T07:36:00.1158835Z * [new branch] gh/malfet/493/orig -> origin/gh/malfet/493/orig 2025-09-07T07:36:00.1159547Z * [new branch] gh/malfet/494/base -> origin/gh/malfet/494/base 2025-09-07T07:36:00.1160080Z * [new branch] gh/malfet/494/head -> origin/gh/malfet/494/head 2025-09-07T07:36:00.1160678Z * [new branch] gh/malfet/494/orig -> origin/gh/malfet/494/orig 2025-09-07T07:36:00.1161602Z * [new branch] gh/malfet/495/base -> origin/gh/malfet/495/base 2025-09-07T07:36:00.1162186Z * [new branch] gh/malfet/495/head -> origin/gh/malfet/495/head 2025-09-07T07:36:00.1162619Z * [new branch] gh/malfet/495/orig -> origin/gh/malfet/495/orig 2025-09-07T07:36:00.1163506Z * [new branch] gh/malfet/496/base -> origin/gh/malfet/496/base 2025-09-07T07:36:00.1163850Z * [new branch] gh/malfet/496/head -> origin/gh/malfet/496/head 2025-09-07T07:36:00.1164571Z * [new branch] gh/malfet/496/orig -> origin/gh/malfet/496/orig 2025-09-07T07:36:00.1165320Z * [new branch] gh/malfet/497/base -> origin/gh/malfet/497/base 2025-09-07T07:36:00.1165854Z * [new branch] gh/malfet/497/head -> origin/gh/malfet/497/head 2025-09-07T07:36:00.1166584Z * [new branch] gh/malfet/497/orig -> origin/gh/malfet/497/orig 2025-09-07T07:36:00.1167255Z * [new branch] gh/malfet/498/base -> origin/gh/malfet/498/base 2025-09-07T07:36:00.1167852Z * [new branch] gh/malfet/498/head -> origin/gh/malfet/498/head 2025-09-07T07:36:00.1168422Z * [new branch] gh/malfet/498/orig -> origin/gh/malfet/498/orig 2025-09-07T07:36:00.1169436Z * [new branch] gh/malfet/499/base -> origin/gh/malfet/499/base 2025-09-07T07:36:00.1169833Z * [new branch] gh/malfet/499/head -> origin/gh/malfet/499/head 2025-09-07T07:36:00.1170419Z * [new branch] gh/malfet/499/orig -> origin/gh/malfet/499/orig 2025-09-07T07:36:00.1171241Z * [new branch] gh/malfet/500/base -> origin/gh/malfet/500/base 2025-09-07T07:36:00.1171560Z * [new branch] gh/malfet/500/head -> origin/gh/malfet/500/head 2025-09-07T07:36:00.1172205Z * [new branch] gh/malfet/500/orig -> origin/gh/malfet/500/orig 2025-09-07T07:36:00.1172959Z * [new branch] gh/malfet/501/base -> origin/gh/malfet/501/base 2025-09-07T07:36:00.1173402Z * [new branch] gh/malfet/501/head -> origin/gh/malfet/501/head 2025-09-07T07:36:00.1174160Z * [new branch] gh/malfet/501/orig -> origin/gh/malfet/501/orig 2025-09-07T07:36:00.1174900Z * [new branch] gh/malfet/502/base -> origin/gh/malfet/502/base 2025-09-07T07:36:00.1175434Z * [new branch] gh/malfet/502/head -> origin/gh/malfet/502/head 2025-09-07T07:36:00.1175976Z * [new branch] gh/malfet/502/orig -> origin/gh/malfet/502/orig 2025-09-07T07:36:00.1176741Z * [new branch] gh/malfet/503/base -> origin/gh/malfet/503/base 2025-09-07T07:36:00.1177298Z * [new branch] gh/malfet/503/head -> origin/gh/malfet/503/head 2025-09-07T07:36:00.1177690Z * [new branch] gh/malfet/503/orig -> origin/gh/malfet/503/orig 2025-09-07T07:36:00.1178506Z * [new branch] gh/malfet/504/base -> origin/gh/malfet/504/base 2025-09-07T07:36:00.1178934Z * [new branch] gh/malfet/504/head -> origin/gh/malfet/504/head 2025-09-07T07:36:00.1179524Z * [new branch] gh/malfet/504/orig -> origin/gh/malfet/504/orig 2025-09-07T07:36:00.1180316Z * [new branch] gh/malfet/505/base -> origin/gh/malfet/505/base 2025-09-07T07:36:00.1180706Z * [new branch] gh/malfet/505/head -> origin/gh/malfet/505/head 2025-09-07T07:36:00.1181529Z * [new branch] gh/malfet/505/orig -> origin/gh/malfet/505/orig 2025-09-07T07:36:00.1182332Z * [new branch] gh/malfet/506/base -> origin/gh/malfet/506/base 2025-09-07T07:36:00.1182611Z * [new branch] gh/malfet/506/head -> origin/gh/malfet/506/head 2025-09-07T07:36:00.1183211Z * [new branch] gh/malfet/506/orig -> origin/gh/malfet/506/orig 2025-09-07T07:36:00.1183963Z * [new branch] gh/malfet/507/base -> origin/gh/malfet/507/base 2025-09-07T07:36:00.1184395Z * [new branch] gh/malfet/507/head -> origin/gh/malfet/507/head 2025-09-07T07:36:00.1185017Z * [new branch] gh/malfet/507/orig -> origin/gh/malfet/507/orig 2025-09-07T07:36:00.1185872Z * [new branch] gh/malfet/508/base -> origin/gh/malfet/508/base 2025-09-07T07:36:00.1186470Z * [new branch] gh/malfet/508/head -> origin/gh/malfet/508/head 2025-09-07T07:36:00.1186896Z * [new branch] gh/malfet/508/orig -> origin/gh/malfet/508/orig 2025-09-07T07:36:00.1187610Z * [new branch] gh/malfet/509/base -> origin/gh/malfet/509/base 2025-09-07T07:36:00.1188029Z * [new branch] gh/malfet/509/head -> origin/gh/malfet/509/head 2025-09-07T07:36:00.1188774Z * [new branch] gh/malfet/509/orig -> origin/gh/malfet/509/orig 2025-09-07T07:36:00.1189534Z * [new branch] gh/malfet/510/base -> origin/gh/malfet/510/base 2025-09-07T07:36:00.1190069Z * [new branch] gh/malfet/510/head -> origin/gh/malfet/510/head 2025-09-07T07:36:00.1190604Z * [new branch] gh/malfet/510/orig -> origin/gh/malfet/510/orig 2025-09-07T07:36:00.1191298Z * [new branch] gh/malfet/511/base -> origin/gh/malfet/511/base 2025-09-07T07:36:00.1191834Z * [new branch] gh/malfet/511/head -> origin/gh/malfet/511/head 2025-09-07T07:36:00.1192519Z * [new branch] gh/malfet/511/orig -> origin/gh/malfet/511/orig 2025-09-07T07:36:00.1193257Z * [new branch] gh/malfet/512/base -> origin/gh/malfet/512/base 2025-09-07T07:36:00.1193798Z * [new branch] gh/malfet/512/head -> origin/gh/malfet/512/head 2025-09-07T07:36:00.1194223Z * [new branch] gh/malfet/512/orig -> origin/gh/malfet/512/orig 2025-09-07T07:36:00.1195027Z * [new branch] gh/malfet/513/base -> origin/gh/malfet/513/base 2025-09-07T07:36:00.1195883Z * [new branch] gh/malfet/513/head -> origin/gh/malfet/513/head 2025-09-07T07:36:00.1196496Z * [new branch] gh/malfet/513/orig -> origin/gh/malfet/513/orig 2025-09-07T07:36:00.1197225Z * [new branch] gh/malfet/64/base -> origin/gh/malfet/64/base 2025-09-07T07:36:00.1197756Z * [new branch] gh/malfet/64/head -> origin/gh/malfet/64/head 2025-09-07T07:36:00.1198691Z * [new branch] gh/manuelcandales/10/base -> origin/gh/manuelcandales/10/base 2025-09-07T07:36:00.1199153Z * [new branch] gh/manuelcandales/10/head -> origin/gh/manuelcandales/10/head 2025-09-07T07:36:00.1199750Z * [new branch] gh/manuelcandales/10/orig -> origin/gh/manuelcandales/10/orig 2025-09-07T07:36:00.1200470Z * [new branch] gh/manuelcandales/11/base -> origin/gh/manuelcandales/11/base 2025-09-07T07:36:00.1200894Z * [new branch] gh/manuelcandales/11/head -> origin/gh/manuelcandales/11/head 2025-09-07T07:36:00.1201491Z * [new branch] gh/manuelcandales/11/orig -> origin/gh/manuelcandales/11/orig 2025-09-07T07:36:00.1202325Z * [new branch] gh/manuelcandales/9/base -> origin/gh/manuelcandales/9/base 2025-09-07T07:36:00.1202874Z * [new branch] gh/manuelcandales/9/head -> origin/gh/manuelcandales/9/head 2025-09-07T07:36:00.1203512Z * [new branch] gh/manuelcandales/9/orig -> origin/gh/manuelcandales/9/orig 2025-09-07T07:36:00.1204667Z * [new branch] gh/markkm/1/base -> origin/gh/markkm/1/base 2025-09-07T07:36:00.1205720Z * [new branch] gh/masnesral/204/base -> origin/gh/masnesral/204/base 2025-09-07T07:36:00.1206348Z * [new branch] gh/masnesral/204/head -> origin/gh/masnesral/204/head 2025-09-07T07:36:00.1206898Z * [new branch] gh/masnesral/204/orig -> origin/gh/masnesral/204/orig 2025-09-07T07:36:00.1207686Z * [new branch] gh/masnesral/235/base -> origin/gh/masnesral/235/base 2025-09-07T07:36:00.1208286Z * [new branch] gh/masnesral/235/head -> origin/gh/masnesral/235/head 2025-09-07T07:36:00.1208800Z * [new branch] gh/masnesral/235/orig -> origin/gh/masnesral/235/orig 2025-09-07T07:36:00.1209618Z * [new branch] gh/masnesral/34/base -> origin/gh/masnesral/34/base 2025-09-07T07:36:00.1210656Z * [new branch] gh/mhorowitz/0/base -> origin/gh/mhorowitz/0/base 2025-09-07T07:36:00.1211199Z * [new branch] gh/mhorowitz/0/head -> origin/gh/mhorowitz/0/head 2025-09-07T07:36:00.1211836Z * [new branch] gh/mhorowitz/1/base -> origin/gh/mhorowitz/1/base 2025-09-07T07:36:00.1212521Z * [new branch] gh/mhorowitz/1/head -> origin/gh/mhorowitz/1/head 2025-09-07T07:36:00.1213169Z * [new branch] gh/mhorowitz/2/base -> origin/gh/mhorowitz/2/base 2025-09-07T07:36:00.1213712Z * [new branch] gh/mhorowitz/2/head -> origin/gh/mhorowitz/2/head 2025-09-07T07:36:00.1214445Z * [new branch] gh/mhorowitz/3/base -> origin/gh/mhorowitz/3/base 2025-09-07T07:36:00.1214803Z * [new branch] gh/mhorowitz/3/head -> origin/gh/mhorowitz/3/head 2025-09-07T07:36:00.1215533Z * [new branch] gh/mhorowitz/4/base -> origin/gh/mhorowitz/4/base 2025-09-07T07:36:00.1215918Z * [new branch] gh/mhorowitz/4/head -> origin/gh/mhorowitz/4/head 2025-09-07T07:36:00.1216760Z * [new branch] gh/mhorowitz/5/base -> origin/gh/mhorowitz/5/base 2025-09-07T07:36:00.1217158Z * [new branch] gh/mhorowitz/5/head -> origin/gh/mhorowitz/5/head 2025-09-07T07:36:00.1217856Z * [new branch] gh/mhorowitz/6/base -> origin/gh/mhorowitz/6/base 2025-09-07T07:36:00.1218252Z * [new branch] gh/mhorowitz/6/head -> origin/gh/mhorowitz/6/head 2025-09-07T07:36:00.1219580Z * [new branch] gh/mikaylagawarecki/234/base -> origin/gh/mikaylagawarecki/234/base 2025-09-07T07:36:00.1220202Z * [new branch] gh/mikaylagawarecki/234/head -> origin/gh/mikaylagawarecki/234/head 2025-09-07T07:36:00.1220814Z * [new branch] gh/mikaylagawarecki/235/base -> origin/gh/mikaylagawarecki/235/base 2025-09-07T07:36:00.1221233Z * [new branch] gh/mikaylagawarecki/235/head -> origin/gh/mikaylagawarecki/235/head 2025-09-07T07:36:00.1222086Z * [new branch] gh/mikaylagawarecki/236/base -> origin/gh/mikaylagawarecki/236/base 2025-09-07T07:36:00.1222484Z * [new branch] gh/mikaylagawarecki/236/head -> origin/gh/mikaylagawarecki/236/head 2025-09-07T07:36:00.1223407Z * [new branch] gh/mikaylagawarecki/237/base -> origin/gh/mikaylagawarecki/237/base 2025-09-07T07:36:00.1223806Z * [new branch] gh/mikaylagawarecki/237/head -> origin/gh/mikaylagawarecki/237/head 2025-09-07T07:36:00.1224615Z * [new branch] gh/mikaylagawarecki/238/base -> origin/gh/mikaylagawarecki/238/base 2025-09-07T07:36:00.1225073Z * [new branch] gh/mikaylagawarecki/238/head -> origin/gh/mikaylagawarecki/238/head 2025-09-07T07:36:00.1225871Z * [new branch] gh/mikaylagawarecki/317/base -> origin/gh/mikaylagawarecki/317/base 2025-09-07T07:36:00.1226440Z * [new branch] gh/mikaylagawarecki/317/head -> origin/gh/mikaylagawarecki/317/head 2025-09-07T07:36:00.1226956Z * [new branch] gh/mikaylagawarecki/317/orig -> origin/gh/mikaylagawarecki/317/orig 2025-09-07T07:36:00.1227757Z * [new branch] gh/mikaylagawarecki/320/base -> origin/gh/mikaylagawarecki/320/base 2025-09-07T07:36:00.1228222Z * [new branch] gh/mikaylagawarecki/320/head -> origin/gh/mikaylagawarecki/320/head 2025-09-07T07:36:00.1228829Z * [new branch] gh/mikaylagawarecki/320/orig -> origin/gh/mikaylagawarecki/320/orig 2025-09-07T07:36:00.1229862Z * [new branch] gh/mikaylagawarecki/329/base -> origin/gh/mikaylagawarecki/329/base 2025-09-07T07:36:00.1230269Z * [new branch] gh/mikaylagawarecki/329/head -> origin/gh/mikaylagawarecki/329/head 2025-09-07T07:36:00.1230943Z * [new branch] gh/mikaylagawarecki/329/orig -> origin/gh/mikaylagawarecki/329/orig 2025-09-07T07:36:00.1231889Z * [new branch] gh/mikaylagawarecki/330/base -> origin/gh/mikaylagawarecki/330/base 2025-09-07T07:36:00.1232323Z * [new branch] gh/mikaylagawarecki/330/head -> origin/gh/mikaylagawarecki/330/head 2025-09-07T07:36:00.1232933Z * [new branch] gh/mikaylagawarecki/330/orig -> origin/gh/mikaylagawarecki/330/orig 2025-09-07T07:36:00.1233697Z * [new branch] gh/mikaylagawarecki/331/base -> origin/gh/mikaylagawarecki/331/base 2025-09-07T07:36:00.1234373Z * [new branch] gh/mikaylagawarecki/331/head -> origin/gh/mikaylagawarecki/331/head 2025-09-07T07:36:00.1234622Z * [new branch] gh/mikaylagawarecki/331/orig -> origin/gh/mikaylagawarecki/331/orig 2025-09-07T07:36:00.1235686Z * [new branch] gh/mikaylagawarecki/332/base -> origin/gh/mikaylagawarecki/332/base 2025-09-07T07:36:00.1236118Z * [new branch] gh/mikaylagawarecki/332/head -> origin/gh/mikaylagawarecki/332/head 2025-09-07T07:36:00.1236685Z * [new branch] gh/mikaylagawarecki/332/orig -> origin/gh/mikaylagawarecki/332/orig 2025-09-07T07:36:00.1237489Z * [new branch] gh/mikaylagawarecki/334/base -> origin/gh/mikaylagawarecki/334/base 2025-09-07T07:36:00.1237910Z * [new branch] gh/mikaylagawarecki/334/head -> origin/gh/mikaylagawarecki/334/head 2025-09-07T07:36:00.1238579Z * [new branch] gh/mikaylagawarecki/334/orig -> origin/gh/mikaylagawarecki/334/orig 2025-09-07T07:36:00.1239353Z * [new branch] gh/mikaylagawarecki/335/base -> origin/gh/mikaylagawarecki/335/base 2025-09-07T07:36:00.1239919Z * [new branch] gh/mikaylagawarecki/335/head -> origin/gh/mikaylagawarecki/335/head 2025-09-07T07:36:00.1240373Z * [new branch] gh/mikaylagawarecki/335/orig -> origin/gh/mikaylagawarecki/335/orig 2025-09-07T07:36:00.1241349Z * [new branch] gh/mikaylagawarecki/336/base -> origin/gh/mikaylagawarecki/336/base 2025-09-07T07:36:00.1241926Z * [new branch] gh/mikaylagawarecki/336/head -> origin/gh/mikaylagawarecki/336/head 2025-09-07T07:36:00.1242489Z * [new branch] gh/mikaylagawarecki/336/orig -> origin/gh/mikaylagawarecki/336/orig 2025-09-07T07:36:00.1243150Z * [new branch] gh/mikaylagawarecki/337/base -> origin/gh/mikaylagawarecki/337/base 2025-09-07T07:36:00.1243583Z * [new branch] gh/mikaylagawarecki/337/head -> origin/gh/mikaylagawarecki/337/head 2025-09-07T07:36:00.1244204Z * [new branch] gh/mikaylagawarecki/337/orig -> origin/gh/mikaylagawarecki/337/orig 2025-09-07T07:36:00.1244984Z * [new branch] gh/mikaylagawarecki/338/base -> origin/gh/mikaylagawarecki/338/base 2025-09-07T07:36:00.1245550Z * [new branch] gh/mikaylagawarecki/338/head -> origin/gh/mikaylagawarecki/338/head 2025-09-07T07:36:00.1246187Z * [new branch] gh/mikaylagawarecki/338/orig -> origin/gh/mikaylagawarecki/338/orig 2025-09-07T07:36:00.1247005Z * [new branch] gh/mikaylagawarecki/339/base -> origin/gh/mikaylagawarecki/339/base 2025-09-07T07:36:00.1247455Z * [new branch] gh/mikaylagawarecki/339/head -> origin/gh/mikaylagawarecki/339/head 2025-09-07T07:36:00.1248123Z * [new branch] gh/mikaylagawarecki/339/orig -> origin/gh/mikaylagawarecki/339/orig 2025-09-07T07:36:00.1249049Z * [new branch] gh/mlazos/1/base -> origin/gh/mlazos/1/base 2025-09-07T07:36:00.1249597Z * [new branch] gh/mlazos/1/head -> origin/gh/mlazos/1/head 2025-09-07T07:36:00.1250212Z * [new branch] gh/mlazos/1/orig -> origin/gh/mlazos/1/orig 2025-09-07T07:36:00.1250955Z * [new branch] gh/mlazos/12/base -> origin/gh/mlazos/12/base 2025-09-07T07:36:00.1251537Z * [new branch] gh/mlazos/12/head -> origin/gh/mlazos/12/head 2025-09-07T07:36:00.1252402Z * [new branch] gh/mlazos/12/orig -> origin/gh/mlazos/12/orig 2025-09-07T07:36:00.1253210Z * [new branch] gh/mlazos/13/base -> origin/gh/mlazos/13/base 2025-09-07T07:36:00.1253893Z * [new branch] gh/mlazos/13/head -> origin/gh/mlazos/13/head 2025-09-07T07:36:00.1254277Z * [new branch] gh/mlazos/13/orig -> origin/gh/mlazos/13/orig 2025-09-07T07:36:00.1255094Z * [new branch] gh/mlazos/14/base -> origin/gh/mlazos/14/base 2025-09-07T07:36:00.1255727Z * [new branch] gh/mlazos/14/head -> origin/gh/mlazos/14/head 2025-09-07T07:36:00.1256083Z * [new branch] gh/mlazos/14/orig -> origin/gh/mlazos/14/orig 2025-09-07T07:36:00.1256950Z * [new branch] gh/mlazos/15/base -> origin/gh/mlazos/15/base 2025-09-07T07:36:00.1257528Z * [new branch] gh/mlazos/15/head -> origin/gh/mlazos/15/head 2025-09-07T07:36:00.1257925Z * [new branch] gh/mlazos/15/orig -> origin/gh/mlazos/15/orig 2025-09-07T07:36:00.1258784Z * [new branch] gh/mlazos/16/base -> origin/gh/mlazos/16/base 2025-09-07T07:36:00.1259353Z * [new branch] gh/mlazos/16/head -> origin/gh/mlazos/16/head 2025-09-07T07:36:00.1260061Z * [new branch] gh/mlazos/16/orig -> origin/gh/mlazos/16/orig 2025-09-07T07:36:00.1260813Z * [new branch] gh/mlazos/17/base -> origin/gh/mlazos/17/base 2025-09-07T07:36:00.1261224Z * [new branch] gh/mlazos/17/head -> origin/gh/mlazos/17/head 2025-09-07T07:36:00.1261826Z * [new branch] gh/mlazos/17/orig -> origin/gh/mlazos/17/orig 2025-09-07T07:36:00.1262874Z * [new branch] gh/mlazos/2/base -> origin/gh/mlazos/2/base 2025-09-07T07:36:00.1263262Z * [new branch] gh/mlazos/2/head -> origin/gh/mlazos/2/head 2025-09-07T07:36:00.1263912Z * [new branch] gh/mlazos/2/orig -> origin/gh/mlazos/2/orig 2025-09-07T07:36:00.1264746Z * [new branch] gh/mlazos/3/base -> origin/gh/mlazos/3/base 2025-09-07T07:36:00.1265082Z * [new branch] gh/mlazos/3/head -> origin/gh/mlazos/3/head 2025-09-07T07:36:00.1265667Z * [new branch] gh/mlazos/3/orig -> origin/gh/mlazos/3/orig 2025-09-07T07:36:00.1266593Z * [new branch] gh/mrmiywj/1/base -> origin/gh/mrmiywj/1/base 2025-09-07T07:36:00.1267319Z * [new branch] gh/mrmiywj/1/head -> origin/gh/mrmiywj/1/head 2025-09-07T07:36:00.1268353Z * [new branch] gh/muchulee8/62/base -> origin/gh/muchulee8/62/base 2025-09-07T07:36:00.1270962Z * [new branch] gh/muchulee8/62/head -> origin/gh/muchulee8/62/head 2025-09-07T07:36:00.1271379Z * [new branch] gh/muchulee8/62/orig -> origin/gh/muchulee8/62/orig 2025-09-07T07:36:00.1272330Z * [new branch] gh/muchulee8/63/base -> origin/gh/muchulee8/63/base 2025-09-07T07:36:00.1272887Z * [new branch] gh/muchulee8/63/head -> origin/gh/muchulee8/63/head 2025-09-07T07:36:00.1273495Z * [new branch] gh/muchulee8/63/orig -> origin/gh/muchulee8/63/orig 2025-09-07T07:36:00.1274316Z * [new branch] gh/muchulee8/64/base -> origin/gh/muchulee8/64/base 2025-09-07T07:36:00.1274722Z * [new branch] gh/muchulee8/64/head -> origin/gh/muchulee8/64/head 2025-09-07T07:36:00.1275378Z * [new branch] gh/muchulee8/64/orig -> origin/gh/muchulee8/64/orig 2025-09-07T07:36:00.1276168Z * [new branch] gh/muchulee8/65/base -> origin/gh/muchulee8/65/base 2025-09-07T07:36:00.1276544Z * [new branch] gh/muchulee8/65/head -> origin/gh/muchulee8/65/head 2025-09-07T07:36:00.1277270Z * [new branch] gh/muchulee8/65/orig -> origin/gh/muchulee8/65/orig 2025-09-07T07:36:00.1278260Z * [new branch] gh/naveenthangudu/1/base -> origin/gh/naveenthangudu/1/base 2025-09-07T07:36:00.1278874Z * [new branch] gh/naveenthangudu/1/head -> origin/gh/naveenthangudu/1/head 2025-09-07T07:36:00.1279452Z * [new branch] gh/naveenthangudu/1/orig -> origin/gh/naveenthangudu/1/orig 2025-09-07T07:36:00.1280223Z * [new branch] gh/naveenthangudu/2/base -> origin/gh/naveenthangudu/2/base 2025-09-07T07:36:00.1280787Z * [new branch] gh/naveenthangudu/2/head -> origin/gh/naveenthangudu/2/head 2025-09-07T07:36:00.1281340Z * [new branch] gh/naveenthangudu/2/orig -> origin/gh/naveenthangudu/2/orig 2025-09-07T07:36:00.1282142Z * [new branch] gh/naveenthangudu/3/base -> origin/gh/naveenthangudu/3/base 2025-09-07T07:36:00.1282606Z * [new branch] gh/naveenthangudu/3/head -> origin/gh/naveenthangudu/3/head 2025-09-07T07:36:00.1283224Z * [new branch] gh/naveenthangudu/3/orig -> origin/gh/naveenthangudu/3/orig 2025-09-07T07:36:00.1283943Z * [new branch] gh/naveenthangudu/4/base -> origin/gh/naveenthangudu/4/base 2025-09-07T07:36:00.1284393Z * [new branch] gh/naveenthangudu/4/head -> origin/gh/naveenthangudu/4/head 2025-09-07T07:36:00.1285128Z * [new branch] gh/naveenthangudu/4/orig -> origin/gh/naveenthangudu/4/orig 2025-09-07T07:36:00.1285937Z * [new branch] gh/naveenthangudu/5/base -> origin/gh/naveenthangudu/5/base 2025-09-07T07:36:00.1286531Z * [new branch] gh/naveenthangudu/5/head -> origin/gh/naveenthangudu/5/head 2025-09-07T07:36:00.1287157Z * [new branch] gh/naveenthangudu/5/orig -> origin/gh/naveenthangudu/5/orig 2025-09-07T07:36:00.1287789Z * [new branch] gh/naveenthangudu/6/base -> origin/gh/naveenthangudu/6/base 2025-09-07T07:36:00.1288231Z * [new branch] gh/naveenthangudu/6/head -> origin/gh/naveenthangudu/6/head 2025-09-07T07:36:00.1288684Z * [new branch] gh/naveenthangudu/6/orig -> origin/gh/naveenthangudu/6/orig 2025-09-07T07:36:00.1289728Z * [new branch] gh/oulgen/35/base -> origin/gh/oulgen/35/base 2025-09-07T07:36:00.1290167Z * [new branch] gh/oulgen/35/head -> origin/gh/oulgen/35/head 2025-09-07T07:36:00.1290829Z * [new branch] gh/oulgen/35/orig -> origin/gh/oulgen/35/orig 2025-09-07T07:36:00.1291617Z * [new branch] gh/oulgen/48/base -> origin/gh/oulgen/48/base 2025-09-07T07:36:00.1292152Z * [new branch] gh/oulgen/48/head -> origin/gh/oulgen/48/head 2025-09-07T07:36:00.1292811Z * [new branch] gh/oulgen/48/orig -> origin/gh/oulgen/48/orig 2025-09-07T07:36:00.1293409Z * [new branch] gh/oulgen/49/base -> origin/gh/oulgen/49/base 2025-09-07T07:36:00.1293954Z * [new branch] gh/oulgen/49/head -> origin/gh/oulgen/49/head 2025-09-07T07:36:00.1294499Z * [new branch] gh/oulgen/49/orig -> origin/gh/oulgen/49/orig 2025-09-07T07:36:00.1295584Z * [new branch] gh/pearu/108/base -> origin/gh/pearu/108/base 2025-09-07T07:36:00.1296170Z * [new branch] gh/pearu/108/head -> origin/gh/pearu/108/head 2025-09-07T07:36:00.1296722Z * [new branch] gh/pearu/108/orig -> origin/gh/pearu/108/orig 2025-09-07T07:36:00.1297525Z * [new branch] gh/pearu/109/base -> origin/gh/pearu/109/base 2025-09-07T07:36:00.1298068Z * [new branch] gh/pearu/109/head -> origin/gh/pearu/109/head 2025-09-07T07:36:00.1298659Z * [new branch] gh/pearu/109/orig -> origin/gh/pearu/109/orig 2025-09-07T07:36:00.1299399Z * [new branch] gh/pearu/110/base -> origin/gh/pearu/110/base 2025-09-07T07:36:00.1299989Z * [new branch] gh/pearu/110/head -> origin/gh/pearu/110/head 2025-09-07T07:36:00.1300543Z * [new branch] gh/pearu/110/orig -> origin/gh/pearu/110/orig 2025-09-07T07:36:00.1301424Z * [new branch] gh/pearu/111/base -> origin/gh/pearu/111/base 2025-09-07T07:36:00.1301799Z * [new branch] gh/pearu/111/head -> origin/gh/pearu/111/head 2025-09-07T07:36:00.1302458Z * [new branch] gh/pearu/111/orig -> origin/gh/pearu/111/orig 2025-09-07T07:36:00.1303130Z * [new branch] gh/pearu/112/base -> origin/gh/pearu/112/base 2025-09-07T07:36:00.1303685Z * [new branch] gh/pearu/112/head -> origin/gh/pearu/112/head 2025-09-07T07:36:00.1304122Z * [new branch] gh/pearu/112/orig -> origin/gh/pearu/112/orig 2025-09-07T07:36:00.1304937Z * [new branch] gh/pearu/113/base -> origin/gh/pearu/113/base 2025-09-07T07:36:00.1305482Z * [new branch] gh/pearu/113/head -> origin/gh/pearu/113/head 2025-09-07T07:36:00.1305919Z * [new branch] gh/pearu/113/orig -> origin/gh/pearu/113/orig 2025-09-07T07:36:00.1306773Z * [new branch] gh/pearu/114/base -> origin/gh/pearu/114/base 2025-09-07T07:36:00.1307695Z * [new branch] gh/pearu/114/head -> origin/gh/pearu/114/head 2025-09-07T07:36:00.1308457Z * [new branch] gh/pearu/114/orig -> origin/gh/pearu/114/orig 2025-09-07T07:36:00.1309081Z * [new branch] gh/pearu/115/base -> origin/gh/pearu/115/base 2025-09-07T07:36:00.1309622Z * [new branch] gh/pearu/115/head -> origin/gh/pearu/115/head 2025-09-07T07:36:00.1310211Z * [new branch] gh/pearu/115/orig -> origin/gh/pearu/115/orig 2025-09-07T07:36:00.1310931Z * [new branch] gh/pearu/116/base -> origin/gh/pearu/116/base 2025-09-07T07:36:00.1311477Z * [new branch] gh/pearu/116/head -> origin/gh/pearu/116/head 2025-09-07T07:36:00.1312332Z * [new branch] gh/pearu/116/orig -> origin/gh/pearu/116/orig 2025-09-07T07:36:00.1313072Z * [new branch] gh/pearu/117/base -> origin/gh/pearu/117/base 2025-09-07T07:36:00.1313458Z * [new branch] gh/pearu/117/head -> origin/gh/pearu/117/head 2025-09-07T07:36:00.1313999Z * [new branch] gh/pearu/117/orig -> origin/gh/pearu/117/orig 2025-09-07T07:36:00.1315044Z * [new branch] gh/pearu/56/base -> origin/gh/pearu/56/base 2025-09-07T07:36:00.1315828Z * [new branch] gh/pearu/56/head -> origin/gh/pearu/56/head 2025-09-07T07:36:00.1316261Z * [new branch] gh/pearu/56/orig -> origin/gh/pearu/56/orig 2025-09-07T07:36:00.1317231Z * [new branch] gh/pearu/97/base -> origin/gh/pearu/97/base 2025-09-07T07:36:00.1317660Z * [new branch] gh/pearu/97/head -> origin/gh/pearu/97/head 2025-09-07T07:36:00.1318264Z * [new branch] gh/pearu/97/orig -> origin/gh/pearu/97/orig 2025-09-07T07:36:00.1319243Z * [new branch] gh/qqaatw/29/base -> origin/gh/qqaatw/29/base 2025-09-07T07:36:00.1319695Z * [new branch] gh/qqaatw/29/head -> origin/gh/qqaatw/29/head 2025-09-07T07:36:00.1320431Z * [new branch] gh/qqaatw/29/orig -> origin/gh/qqaatw/29/orig 2025-09-07T07:36:00.1321216Z * [new branch] gh/raymo/refresh-script -> origin/gh/raymo/refresh-script 2025-09-07T07:36:00.1322058Z * [new branch] gh/rec/141/base -> origin/gh/rec/141/base 2025-09-07T07:36:00.1322986Z * [new branch] gh/rec/141/head -> origin/gh/rec/141/head 2025-09-07T07:36:00.1323721Z * [new branch] gh/rec/153/base -> origin/gh/rec/153/base 2025-09-07T07:36:00.1324252Z * [new branch] gh/rec/153/head -> origin/gh/rec/153/head 2025-09-07T07:36:00.1324794Z * [new branch] gh/rec/153/orig -> origin/gh/rec/153/orig 2025-09-07T07:36:00.1325572Z * [new branch] gh/rec/154/base -> origin/gh/rec/154/base 2025-09-07T07:36:00.1326109Z * [new branch] gh/rec/154/head -> origin/gh/rec/154/head 2025-09-07T07:36:00.1326651Z * [new branch] gh/rec/154/orig -> origin/gh/rec/154/orig 2025-09-07T07:36:00.1327350Z * [new branch] gh/rec/156/base -> origin/gh/rec/156/base 2025-09-07T07:36:00.1327879Z * [new branch] gh/rec/156/head -> origin/gh/rec/156/head 2025-09-07T07:36:00.1328411Z * [new branch] gh/rec/156/orig -> origin/gh/rec/156/orig 2025-09-07T07:36:00.1329163Z * [new branch] gh/rec/160/base -> origin/gh/rec/160/base 2025-09-07T07:36:00.1329685Z * [new branch] gh/rec/160/head -> origin/gh/rec/160/head 2025-09-07T07:36:00.1330403Z * [new branch] gh/rec/160/orig -> origin/gh/rec/160/orig 2025-09-07T07:36:00.1331101Z * [new branch] gh/rec/162/base -> origin/gh/rec/162/base 2025-09-07T07:36:00.1331647Z * [new branch] gh/rec/162/head -> origin/gh/rec/162/head 2025-09-07T07:36:00.1332199Z * [new branch] gh/rec/162/orig -> origin/gh/rec/162/orig 2025-09-07T07:36:00.1333191Z * [new branch] gh/rec/163/base -> origin/gh/rec/163/base 2025-09-07T07:36:00.1333578Z * [new branch] gh/rec/163/head -> origin/gh/rec/163/head 2025-09-07T07:36:00.1334173Z * [new branch] gh/rec/163/orig -> origin/gh/rec/163/orig 2025-09-07T07:36:00.1334924Z * [new branch] gh/rec/164/base -> origin/gh/rec/164/base 2025-09-07T07:36:00.1335452Z * [new branch] gh/rec/164/head -> origin/gh/rec/164/head 2025-09-07T07:36:00.1335992Z * [new branch] gh/rec/164/orig -> origin/gh/rec/164/orig 2025-09-07T07:36:00.1336736Z * [new branch] gh/rec/165/base -> origin/gh/rec/165/base 2025-09-07T07:36:00.1337274Z * [new branch] gh/rec/165/head -> origin/gh/rec/165/head 2025-09-07T07:36:00.1337846Z * [new branch] gh/rec/165/orig -> origin/gh/rec/165/orig 2025-09-07T07:36:00.1338537Z * [new branch] gh/rec/166/base -> origin/gh/rec/166/base 2025-09-07T07:36:00.1339064Z * [new branch] gh/rec/166/head -> origin/gh/rec/166/head 2025-09-07T07:36:00.1339684Z * [new branch] gh/rec/166/orig -> origin/gh/rec/166/orig 2025-09-07T07:36:00.1340675Z * [new branch] gh/robert-hardwick/1/base -> origin/gh/robert-hardwick/1/base 2025-09-07T07:36:00.1341088Z * [new branch] gh/robert-hardwick/1/head -> origin/gh/robert-hardwick/1/head 2025-09-07T07:36:00.1341709Z * [new branch] gh/robert-hardwick/1/orig -> origin/gh/robert-hardwick/1/orig 2025-09-07T07:36:00.1342456Z * [new branch] gh/robert-hardwick/2/base -> origin/gh/robert-hardwick/2/base 2025-09-07T07:36:00.1342919Z * [new branch] gh/robert-hardwick/2/head -> origin/gh/robert-hardwick/2/head 2025-09-07T07:36:00.1343542Z * [new branch] gh/robert-hardwick/2/orig -> origin/gh/robert-hardwick/2/orig 2025-09-07T07:36:00.1344242Z * [new branch] gh/robert-hardwick/3/base -> origin/gh/robert-hardwick/3/base 2025-09-07T07:36:00.1344893Z * [new branch] gh/robert-hardwick/3/head -> origin/gh/robert-hardwick/3/head 2025-09-07T07:36:00.1345339Z * [new branch] gh/robert-hardwick/3/orig -> origin/gh/robert-hardwick/3/orig 2025-09-07T07:36:00.1346138Z * [new branch] gh/robert-hardwick/4/base -> origin/gh/robert-hardwick/4/base 2025-09-07T07:36:00.1346573Z * [new branch] gh/robert-hardwick/4/head -> origin/gh/robert-hardwick/4/head 2025-09-07T07:36:00.1347180Z * [new branch] gh/robert-hardwick/4/orig -> origin/gh/robert-hardwick/4/orig 2025-09-07T07:36:00.1348204Z * [new branch] gh/rtimpe/1/base -> origin/gh/rtimpe/1/base 2025-09-07T07:36:00.1348528Z * [new branch] gh/rtimpe/1/head -> origin/gh/rtimpe/1/head 2025-09-07T07:36:00.1349773Z * [new branch] gh/rtimpe/10/base -> origin/gh/rtimpe/10/base 2025-09-07T07:36:00.1350452Z * [new branch] gh/rtimpe/10/head -> origin/gh/rtimpe/10/head 2025-09-07T07:36:00.1350852Z * [new branch] gh/rtimpe/10/orig -> origin/gh/rtimpe/10/orig 2025-09-07T07:36:00.1351842Z * [new branch] gh/rtimpe/11/base -> origin/gh/rtimpe/11/base 2025-09-07T07:36:00.1352220Z * [new branch] gh/rtimpe/11/head -> origin/gh/rtimpe/11/head 2025-09-07T07:36:00.1352812Z * [new branch] gh/rtimpe/11/orig -> origin/gh/rtimpe/11/orig 2025-09-07T07:36:00.1353652Z * [new branch] gh/rtimpe/12/base -> origin/gh/rtimpe/12/base 2025-09-07T07:36:00.1354042Z * [new branch] gh/rtimpe/12/head -> origin/gh/rtimpe/12/head 2025-09-07T07:36:00.1354624Z * [new branch] gh/rtimpe/12/orig -> origin/gh/rtimpe/12/orig 2025-09-07T07:36:00.1355366Z * [new branch] gh/rtimpe/13/base -> origin/gh/rtimpe/13/base 2025-09-07T07:36:00.1355792Z * [new branch] gh/rtimpe/13/head -> origin/gh/rtimpe/13/head 2025-09-07T07:36:00.1356475Z * [new branch] gh/rtimpe/13/orig -> origin/gh/rtimpe/13/orig 2025-09-07T07:36:00.1357267Z * [new branch] gh/rtimpe/14/base -> origin/gh/rtimpe/14/base 2025-09-07T07:36:00.1357588Z * [new branch] gh/rtimpe/14/head -> origin/gh/rtimpe/14/head 2025-09-07T07:36:00.1358277Z * [new branch] gh/rtimpe/14/orig -> origin/gh/rtimpe/14/orig 2025-09-07T07:36:00.1359122Z * [new branch] gh/rtimpe/15/base -> origin/gh/rtimpe/15/base 2025-09-07T07:36:00.1359509Z * [new branch] gh/rtimpe/15/head -> origin/gh/rtimpe/15/head 2025-09-07T07:36:00.1360151Z * [new branch] gh/rtimpe/15/orig -> origin/gh/rtimpe/15/orig 2025-09-07T07:36:00.1360854Z * [new branch] gh/rtimpe/2/base -> origin/gh/rtimpe/2/base 2025-09-07T07:36:00.1361238Z * [new branch] gh/rtimpe/2/head -> origin/gh/rtimpe/2/head 2025-09-07T07:36:00.1362118Z * [new branch] gh/rtimpe/3/base -> origin/gh/rtimpe/3/base 2025-09-07T07:36:00.1362434Z * [new branch] gh/rtimpe/3/head -> origin/gh/rtimpe/3/head 2025-09-07T07:36:00.1363366Z * [new branch] gh/rtimpe/4/base -> origin/gh/rtimpe/4/base 2025-09-07T07:36:00.1363677Z * [new branch] gh/rtimpe/4/head -> origin/gh/rtimpe/4/head 2025-09-07T07:36:00.1364644Z * [new branch] gh/rtimpe/9/base -> origin/gh/rtimpe/9/base 2025-09-07T07:36:00.1365048Z * [new branch] gh/rtimpe/9/head -> origin/gh/rtimpe/9/head 2025-09-07T07:36:00.1365669Z * [new branch] gh/rtimpe/9/orig -> origin/gh/rtimpe/9/orig 2025-09-07T07:36:00.1366982Z * [new branch] gh/ruisizhang123/1/base -> origin/gh/ruisizhang123/1/base 2025-09-07T07:36:00.1367405Z * [new branch] gh/ruisizhang123/1/head -> origin/gh/ruisizhang123/1/head 2025-09-07T07:36:00.1368110Z * [new branch] gh/ruisizhang123/1/orig -> origin/gh/ruisizhang123/1/orig 2025-09-07T07:36:00.1368874Z * [new branch] gh/ruisizhang123/4/base -> origin/gh/ruisizhang123/4/base 2025-09-07T07:36:00.1369329Z * [new branch] gh/ruisizhang123/4/head -> origin/gh/ruisizhang123/4/head 2025-09-07T07:36:00.1369945Z * [new branch] gh/ruisizhang123/4/orig -> origin/gh/ruisizhang123/4/orig 2025-09-07T07:36:00.1371074Z * [new branch] gh/ruisizhang123/5/base -> origin/gh/ruisizhang123/5/base 2025-09-07T07:36:00.1371491Z * [new branch] gh/ruisizhang123/5/head -> origin/gh/ruisizhang123/5/head 2025-09-07T07:36:00.1372097Z * [new branch] gh/ruisizhang123/5/orig -> origin/gh/ruisizhang123/5/orig 2025-09-07T07:36:00.1372866Z * [new branch] gh/ruisizhang123/6/base -> origin/gh/ruisizhang123/6/base 2025-09-07T07:36:00.1373272Z * [new branch] gh/ruisizhang123/6/head -> origin/gh/ruisizhang123/6/head 2025-09-07T07:36:00.1373946Z * [new branch] gh/ruisizhang123/6/orig -> origin/gh/ruisizhang123/6/orig 2025-09-07T07:36:00.1374642Z * [new branch] gh/ruisizhang123/7/base -> origin/gh/ruisizhang123/7/base 2025-09-07T07:36:00.1375207Z * [new branch] gh/ruisizhang123/7/head -> origin/gh/ruisizhang123/7/head 2025-09-07T07:36:00.1375772Z * [new branch] gh/ruisizhang123/7/orig -> origin/gh/ruisizhang123/7/orig 2025-09-07T07:36:00.1376427Z * [new branch] gh/ruisizhang123/8/base -> origin/gh/ruisizhang123/8/base 2025-09-07T07:36:00.1377019Z * [new branch] gh/ruisizhang123/8/head -> origin/gh/ruisizhang123/8/head 2025-09-07T07:36:00.1377575Z * [new branch] gh/ruisizhang123/8/orig -> origin/gh/ruisizhang123/8/orig 2025-09-07T07:36:00.1378303Z * [new branch] gh/ruisizhang123/9/base -> origin/gh/ruisizhang123/9/base 2025-09-07T07:36:00.1378754Z * [new branch] gh/ruisizhang123/9/head -> origin/gh/ruisizhang123/9/head 2025-09-07T07:36:00.1379648Z * [new branch] gh/ruisizhang123/9/orig -> origin/gh/ruisizhang123/9/orig 2025-09-07T07:36:00.1380686Z * [new branch] gh/sarckk/2/base -> origin/gh/sarckk/2/base 2025-09-07T07:36:00.1381067Z * [new branch] gh/sarckk/2/head -> origin/gh/sarckk/2/head 2025-09-07T07:36:00.1381816Z * [new branch] gh/sarckk/2/orig -> origin/gh/sarckk/2/orig 2025-09-07T07:36:00.1382734Z * [new branch] gh/seemethere/35/base -> origin/gh/seemethere/35/base 2025-09-07T07:36:00.1383467Z * [new branch] gh/seemethere/35/head -> origin/gh/seemethere/35/head 2025-09-07T07:36:00.1384087Z * [new branch] gh/seemethere/35/orig -> origin/gh/seemethere/35/orig 2025-09-07T07:36:00.1384757Z * [new branch] gh/seemethere/37/base -> origin/gh/seemethere/37/base 2025-09-07T07:36:00.1385309Z * [new branch] gh/seemethere/37/head -> origin/gh/seemethere/37/head 2025-09-07T07:36:00.1385754Z * [new branch] gh/seemethere/37/orig -> origin/gh/seemethere/37/orig 2025-09-07T07:36:00.1386646Z * [new branch] gh/seemethere/43/base -> origin/gh/seemethere/43/base 2025-09-07T07:36:00.1386952Z * [new branch] gh/seemethere/43/head -> origin/gh/seemethere/43/head 2025-09-07T07:36:00.1387663Z * [new branch] gh/seemethere/43/orig -> origin/gh/seemethere/43/orig 2025-09-07T07:36:00.1388414Z * [new branch] gh/seemethere/44/base -> origin/gh/seemethere/44/base 2025-09-07T07:36:00.1388967Z * [new branch] gh/seemethere/44/head -> origin/gh/seemethere/44/head 2025-09-07T07:36:00.1389515Z * [new branch] gh/seemethere/44/orig -> origin/gh/seemethere/44/orig 2025-09-07T07:36:00.1390195Z * [new branch] gh/seemethere/48/base -> origin/gh/seemethere/48/base 2025-09-07T07:36:00.1390851Z * [new branch] gh/seemethere/48/head -> origin/gh/seemethere/48/head 2025-09-07T07:36:00.1391249Z * [new branch] gh/seemethere/48/orig -> origin/gh/seemethere/48/orig 2025-09-07T07:36:00.1392095Z * [new branch] gh/seemethere/49/base -> origin/gh/seemethere/49/base 2025-09-07T07:36:00.1392542Z * [new branch] gh/seemethere/49/head -> origin/gh/seemethere/49/head 2025-09-07T07:36:00.1393178Z * [new branch] gh/seemethere/49/orig -> origin/gh/seemethere/49/orig 2025-09-07T07:36:00.1393945Z * [new branch] gh/seemethere/52/base -> origin/gh/seemethere/52/base 2025-09-07T07:36:00.1394342Z * [new branch] gh/seemethere/52/head -> origin/gh/seemethere/52/head 2025-09-07T07:36:00.1394952Z * [new branch] gh/seemethere/52/orig -> origin/gh/seemethere/52/orig 2025-09-07T07:36:00.1395677Z * [new branch] gh/seemethere/53/base -> origin/gh/seemethere/53/base 2025-09-07T07:36:00.1396283Z * [new branch] gh/seemethere/53/head -> origin/gh/seemethere/53/head 2025-09-07T07:36:00.1396889Z * [new branch] gh/seemethere/53/orig -> origin/gh/seemethere/53/orig 2025-09-07T07:36:00.1397613Z * [new branch] gh/seemethere/54/base -> origin/gh/seemethere/54/base 2025-09-07T07:36:00.1398052Z * [new branch] gh/seemethere/54/head -> origin/gh/seemethere/54/head 2025-09-07T07:36:00.1398694Z * [new branch] gh/seemethere/54/orig -> origin/gh/seemethere/54/orig 2025-09-07T07:36:00.1399398Z * [new branch] gh/seemethere/55/base -> origin/gh/seemethere/55/base 2025-09-07T07:36:00.1399812Z * [new branch] gh/seemethere/55/head -> origin/gh/seemethere/55/head 2025-09-07T07:36:00.1400351Z * [new branch] gh/seemethere/55/orig -> origin/gh/seemethere/55/orig 2025-09-07T07:36:00.1401163Z * [new branch] gh/seemethere/56/base -> origin/gh/seemethere/56/base 2025-09-07T07:36:00.1401556Z * [new branch] gh/seemethere/56/head -> origin/gh/seemethere/56/head 2025-09-07T07:36:00.1402180Z * [new branch] gh/seemethere/56/orig -> origin/gh/seemethere/56/orig 2025-09-07T07:36:00.1403260Z * [new branch] gh/seemethere/57/base -> origin/gh/seemethere/57/base 2025-09-07T07:36:00.1403684Z * [new branch] gh/seemethere/57/head -> origin/gh/seemethere/57/head 2025-09-07T07:36:00.1404315Z * [new branch] gh/seemethere/57/orig -> origin/gh/seemethere/57/orig 2025-09-07T07:36:00.1405021Z * [new branch] gh/seemethere/58/base -> origin/gh/seemethere/58/base 2025-09-07T07:36:00.1405682Z * [new branch] gh/seemethere/58/head -> origin/gh/seemethere/58/head 2025-09-07T07:36:00.1406091Z * [new branch] gh/seemethere/58/orig -> origin/gh/seemethere/58/orig 2025-09-07T07:36:00.1406853Z * [new branch] gh/seemethere/59/base -> origin/gh/seemethere/59/base 2025-09-07T07:36:00.1407357Z * [new branch] gh/seemethere/59/head -> origin/gh/seemethere/59/head 2025-09-07T07:36:00.1407811Z * [new branch] gh/seemethere/59/orig -> origin/gh/seemethere/59/orig 2025-09-07T07:36:00.1408594Z * [new branch] gh/seemethere/60/base -> origin/gh/seemethere/60/base 2025-09-07T07:36:00.1409212Z * [new branch] gh/seemethere/60/head -> origin/gh/seemethere/60/head 2025-09-07T07:36:00.1409697Z * [new branch] gh/seemethere/60/orig -> origin/gh/seemethere/60/orig 2025-09-07T07:36:00.1410475Z * [new branch] gh/seemethere/61/base -> origin/gh/seemethere/61/base 2025-09-07T07:36:00.1411083Z * [new branch] gh/seemethere/61/head -> origin/gh/seemethere/61/head 2025-09-07T07:36:00.1411623Z * [new branch] gh/seemethere/61/orig -> origin/gh/seemethere/61/orig 2025-09-07T07:36:00.1412335Z * [new branch] gh/seemethere/62/base -> origin/gh/seemethere/62/base 2025-09-07T07:36:00.1412769Z * [new branch] gh/seemethere/62/head -> origin/gh/seemethere/62/head 2025-09-07T07:36:00.1413387Z * [new branch] gh/seemethere/62/orig -> origin/gh/seemethere/62/orig 2025-09-07T07:36:00.1414210Z * [new branch] gh/seemethere/63/base -> origin/gh/seemethere/63/base 2025-09-07T07:36:00.1414664Z * [new branch] gh/seemethere/63/head -> origin/gh/seemethere/63/head 2025-09-07T07:36:00.1415274Z * [new branch] gh/seemethere/63/orig -> origin/gh/seemethere/63/orig 2025-09-07T07:36:00.1416419Z * [new branch] gh/shunting314/145/base -> origin/gh/shunting314/145/base 2025-09-07T07:36:00.1416869Z * [new branch] gh/shunting314/145/head -> origin/gh/shunting314/145/head 2025-09-07T07:36:00.1417562Z * [new branch] gh/shunting314/145/orig -> origin/gh/shunting314/145/orig 2025-09-07T07:36:00.1418616Z * [new branch] gh/shunting314/176/base -> origin/gh/shunting314/176/base 2025-09-07T07:36:00.1419224Z * [new branch] gh/shunting314/176/head -> origin/gh/shunting314/176/head 2025-09-07T07:36:00.1419829Z * [new branch] gh/shunting314/176/orig -> origin/gh/shunting314/176/orig 2025-09-07T07:36:00.1420561Z * [new branch] gh/shunting314/211/base -> origin/gh/shunting314/211/base 2025-09-07T07:36:00.1421146Z * [new branch] gh/shunting314/211/head -> origin/gh/shunting314/211/head 2025-09-07T07:36:00.1421539Z * [new branch] gh/shunting314/211/orig -> origin/gh/shunting314/211/orig 2025-09-07T07:36:00.1422283Z * [new branch] gh/shunting314/212/base -> origin/gh/shunting314/212/base 2025-09-07T07:36:00.1422728Z * [new branch] gh/shunting314/212/head -> origin/gh/shunting314/212/head 2025-09-07T07:36:00.1423417Z * [new branch] gh/shunting314/212/orig -> origin/gh/shunting314/212/orig 2025-09-07T07:36:00.1424431Z * [new branch] gh/shunting314/213/base -> origin/gh/shunting314/213/base 2025-09-07T07:36:00.1425044Z * [new branch] gh/shunting314/213/head -> origin/gh/shunting314/213/head 2025-09-07T07:36:00.1425592Z * [new branch] gh/shunting314/213/orig -> origin/gh/shunting314/213/orig 2025-09-07T07:36:00.1426547Z * [new branch] gh/shunting314/214/base -> origin/gh/shunting314/214/base 2025-09-07T07:36:00.1426914Z * [new branch] gh/shunting314/214/head -> origin/gh/shunting314/214/head 2025-09-07T07:36:00.1427536Z * [new branch] gh/shunting314/214/orig -> origin/gh/shunting314/214/orig 2025-09-07T07:36:00.1428449Z * [new branch] gh/shunting314/215/base -> origin/gh/shunting314/215/base 2025-09-07T07:36:00.1428848Z * [new branch] gh/shunting314/215/head -> origin/gh/shunting314/215/head 2025-09-07T07:36:00.1429477Z * [new branch] gh/shunting314/215/orig -> origin/gh/shunting314/215/orig 2025-09-07T07:36:00.1430243Z * [new branch] gh/shunting314/216/base -> origin/gh/shunting314/216/base 2025-09-07T07:36:00.1430578Z * [new branch] gh/shunting314/216/head -> origin/gh/shunting314/216/head 2025-09-07T07:36:00.1431177Z * [new branch] gh/shunting314/216/orig -> origin/gh/shunting314/216/orig 2025-09-07T07:36:00.1432116Z * [new branch] gh/shunting314/217/base -> origin/gh/shunting314/217/base 2025-09-07T07:36:00.1432399Z * [new branch] gh/shunting314/217/head -> origin/gh/shunting314/217/head 2025-09-07T07:36:00.1432957Z * [new branch] gh/shunting314/217/orig -> origin/gh/shunting314/217/orig 2025-09-07T07:36:00.1433915Z * [new branch] gh/shunting314/218/base -> origin/gh/shunting314/218/base 2025-09-07T07:36:00.1434554Z * [new branch] gh/shunting314/218/head -> origin/gh/shunting314/218/head 2025-09-07T07:36:00.1435110Z * [new branch] gh/shunting314/218/orig -> origin/gh/shunting314/218/orig 2025-09-07T07:36:00.1436130Z * [new branch] gh/shunting314/219/base -> origin/gh/shunting314/219/base 2025-09-07T07:36:00.1436543Z * [new branch] gh/shunting314/219/head -> origin/gh/shunting314/219/head 2025-09-07T07:36:00.1437464Z * [new branch] gh/shunting314/219/orig -> origin/gh/shunting314/219/orig 2025-09-07T07:36:00.1438397Z * [new branch] gh/shunting314/220/base -> origin/gh/shunting314/220/base 2025-09-07T07:36:00.1439012Z * [new branch] gh/shunting314/220/head -> origin/gh/shunting314/220/head 2025-09-07T07:36:00.1439431Z * [new branch] gh/shunting314/220/orig -> origin/gh/shunting314/220/orig 2025-09-07T07:36:00.1440214Z * [new branch] gh/shunting314/221/base -> origin/gh/shunting314/221/base 2025-09-07T07:36:00.1440636Z * [new branch] gh/shunting314/221/head -> origin/gh/shunting314/221/head 2025-09-07T07:36:00.1441243Z * [new branch] gh/shunting314/221/orig -> origin/gh/shunting314/221/orig 2025-09-07T07:36:00.1441992Z * [new branch] gh/shunting314/222/base -> origin/gh/shunting314/222/base 2025-09-07T07:36:00.1442422Z * [new branch] gh/shunting314/222/head -> origin/gh/shunting314/222/head 2025-09-07T07:36:00.1443049Z * [new branch] gh/shunting314/222/orig -> origin/gh/shunting314/222/orig 2025-09-07T07:36:00.1443707Z * [new branch] gh/shunting314/223/base -> origin/gh/shunting314/223/base 2025-09-07T07:36:00.1444380Z * [new branch] gh/shunting314/223/head -> origin/gh/shunting314/223/head 2025-09-07T07:36:00.1444934Z * [new branch] gh/shunting314/223/orig -> origin/gh/shunting314/223/orig 2025-09-07T07:36:00.1445862Z * [new branch] gh/silverguo/1/base -> origin/gh/silverguo/1/base 2025-09-07T07:36:00.1446416Z * [new branch] gh/silverguo/1/head -> origin/gh/silverguo/1/head 2025-09-07T07:36:00.1447049Z * [new branch] gh/silverguo/2/base -> origin/gh/silverguo/2/base 2025-09-07T07:36:00.1447463Z * [new branch] gh/silverguo/2/head -> origin/gh/silverguo/2/head 2025-09-07T07:36:00.1448165Z * [new branch] gh/silverguo/3/base -> origin/gh/silverguo/3/base 2025-09-07T07:36:00.1448901Z * [new branch] gh/silverguo/3/head -> origin/gh/silverguo/3/head 2025-09-07T07:36:00.1449449Z * [new branch] gh/silverguo/4/base -> origin/gh/silverguo/4/base 2025-09-07T07:36:00.1449884Z * [new branch] gh/silverguo/4/head -> origin/gh/silverguo/4/head 2025-09-07T07:36:00.1450864Z * [new branch] gh/sinhaanhsul/1/base -> origin/gh/sinhaanhsul/1/base 2025-09-07T07:36:00.1451412Z * [new branch] gh/sinhaanhsul/1/head -> origin/gh/sinhaanhsul/1/head 2025-09-07T07:36:00.1452348Z * [new branch] gh/skarjala/17/base -> origin/gh/skarjala/17/base 2025-09-07T07:36:00.1452783Z * [new branch] gh/skarjala/17/head -> origin/gh/skarjala/17/head 2025-09-07T07:36:00.1453586Z * [new branch] gh/skarjala/17/orig -> origin/gh/skarjala/17/orig 2025-09-07T07:36:00.1454345Z * [new branch] gh/skarjala/18/base -> origin/gh/skarjala/18/base 2025-09-07T07:36:00.1454830Z * [new branch] gh/skarjala/18/head -> origin/gh/skarjala/18/head 2025-09-07T07:36:00.1455465Z * [new branch] gh/skarjala/18/orig -> origin/gh/skarjala/18/orig 2025-09-07T07:36:00.1456189Z * [new branch] gh/skarjala/19/base -> origin/gh/skarjala/19/base 2025-09-07T07:36:00.1456840Z * [new branch] gh/skarjala/19/head -> origin/gh/skarjala/19/head 2025-09-07T07:36:00.1457180Z * [new branch] gh/skarjala/19/orig -> origin/gh/skarjala/19/orig 2025-09-07T07:36:00.1458118Z * [new branch] gh/slayton58/1/base -> origin/gh/slayton58/1/base 2025-09-07T07:36:00.1458967Z * [new branch] gh/slayton58/1/head -> origin/gh/slayton58/1/head 2025-09-07T07:36:00.1459514Z * [new branch] gh/slayton58/1/orig -> origin/gh/slayton58/1/orig 2025-09-07T07:36:00.1460177Z * [new branch] gh/slayton58/2/base -> origin/gh/slayton58/2/base 2025-09-07T07:36:00.1460634Z * [new branch] gh/slayton58/2/head -> origin/gh/slayton58/2/head 2025-09-07T07:36:00.1461242Z * [new branch] gh/slayton58/2/orig -> origin/gh/slayton58/2/orig 2025-09-07T07:36:00.1461932Z * [new branch] gh/slayton58/3/base -> origin/gh/slayton58/3/base 2025-09-07T07:36:00.1462526Z * [new branch] gh/slayton58/3/head -> origin/gh/slayton58/3/head 2025-09-07T07:36:00.1463251Z * [new branch] gh/slayton58/3/orig -> origin/gh/slayton58/3/orig 2025-09-07T07:36:00.1463945Z * [new branch] gh/slayton58/4/base -> origin/gh/slayton58/4/base 2025-09-07T07:36:00.1464369Z * [new branch] gh/slayton58/4/head -> origin/gh/slayton58/4/head 2025-09-07T07:36:00.1464963Z * [new branch] gh/slayton58/4/orig -> origin/gh/slayton58/4/orig 2025-09-07T07:36:00.1465618Z * [new branch] gh/slayton58/5/base -> origin/gh/slayton58/5/base 2025-09-07T07:36:00.1466060Z * [new branch] gh/slayton58/5/head -> origin/gh/slayton58/5/head 2025-09-07T07:36:00.1466951Z * [new branch] gh/slayton58/5/orig -> origin/gh/slayton58/5/orig 2025-09-07T07:36:00.1468107Z * [new branch] gh/soulitzer/269/base -> origin/gh/soulitzer/269/base 2025-09-07T07:36:00.1468468Z * [new branch] gh/soulitzer/269/head -> origin/gh/soulitzer/269/head 2025-09-07T07:36:00.1469104Z * [new branch] gh/soulitzer/269/orig -> origin/gh/soulitzer/269/orig 2025-09-07T07:36:00.1469974Z * [new branch] gh/soulitzer/276/base -> origin/gh/soulitzer/276/base 2025-09-07T07:36:00.1470529Z * [new branch] gh/soulitzer/276/head -> origin/gh/soulitzer/276/head 2025-09-07T07:36:00.1471072Z * [new branch] gh/soulitzer/276/orig -> origin/gh/soulitzer/276/orig 2025-09-07T07:36:00.1472067Z * [new branch] gh/soulitzer/287/base -> origin/gh/soulitzer/287/base 2025-09-07T07:36:00.1472508Z * [new branch] gh/soulitzer/287/head -> origin/gh/soulitzer/287/head 2025-09-07T07:36:00.1473243Z * [new branch] gh/soulitzer/287/orig -> origin/gh/soulitzer/287/orig 2025-09-07T07:36:00.1474084Z * [new branch] gh/soulitzer/296/base -> origin/gh/soulitzer/296/base 2025-09-07T07:36:00.1474527Z * [new branch] gh/soulitzer/296/head -> origin/gh/soulitzer/296/head 2025-09-07T07:36:00.1475163Z * [new branch] gh/soulitzer/296/orig -> origin/gh/soulitzer/296/orig 2025-09-07T07:36:00.1475882Z * [new branch] gh/soulitzer/299/base -> origin/gh/soulitzer/299/base 2025-09-07T07:36:00.1476419Z * [new branch] gh/soulitzer/299/head -> origin/gh/soulitzer/299/head 2025-09-07T07:36:00.1476908Z * [new branch] gh/soulitzer/299/orig -> origin/gh/soulitzer/299/orig 2025-09-07T07:36:00.1477869Z * [new branch] gh/soulitzer/300/base -> origin/gh/soulitzer/300/base 2025-09-07T07:36:00.1478416Z * [new branch] gh/soulitzer/300/head -> origin/gh/soulitzer/300/head 2025-09-07T07:36:00.1478849Z * [new branch] gh/soulitzer/300/orig -> origin/gh/soulitzer/300/orig 2025-09-07T07:36:00.1479748Z * [new branch] gh/soulitzer/301/base -> origin/gh/soulitzer/301/base 2025-09-07T07:36:00.1480342Z * [new branch] gh/soulitzer/301/head -> origin/gh/soulitzer/301/head 2025-09-07T07:36:00.1480904Z * [new branch] gh/soulitzer/301/orig -> origin/gh/soulitzer/301/orig 2025-09-07T07:36:00.1481631Z * [new branch] gh/soulitzer/313/base -> origin/gh/soulitzer/313/base 2025-09-07T07:36:00.1482184Z * [new branch] gh/soulitzer/313/head -> origin/gh/soulitzer/313/head 2025-09-07T07:36:00.1482837Z * [new branch] gh/soulitzer/313/orig -> origin/gh/soulitzer/313/orig 2025-09-07T07:36:00.1483623Z * [new branch] gh/soulitzer/319/base -> origin/gh/soulitzer/319/base 2025-09-07T07:36:00.1484180Z * [new branch] gh/soulitzer/319/head -> origin/gh/soulitzer/319/head 2025-09-07T07:36:00.1484603Z * [new branch] gh/soulitzer/319/orig -> origin/gh/soulitzer/319/orig 2025-09-07T07:36:00.1485927Z * [new branch] gh/soulitzer/320/base -> origin/gh/soulitzer/320/base 2025-09-07T07:36:00.1486491Z * [new branch] gh/soulitzer/320/head -> origin/gh/soulitzer/320/head 2025-09-07T07:36:00.1486851Z * [new branch] gh/soulitzer/320/orig -> origin/gh/soulitzer/320/orig 2025-09-07T07:36:00.1487772Z * [new branch] gh/soulitzer/336/base -> origin/gh/soulitzer/336/base 2025-09-07T07:36:00.1488175Z * [new branch] gh/soulitzer/336/head -> origin/gh/soulitzer/336/head 2025-09-07T07:36:00.1488815Z * [new branch] gh/soulitzer/336/orig -> origin/gh/soulitzer/336/orig 2025-09-07T07:36:00.1489506Z * [new branch] gh/soulitzer/347/base -> origin/gh/soulitzer/347/base 2025-09-07T07:36:00.1489903Z * [new branch] gh/soulitzer/347/head -> origin/gh/soulitzer/347/head 2025-09-07T07:36:00.1490465Z * [new branch] gh/soulitzer/347/orig -> origin/gh/soulitzer/347/orig 2025-09-07T07:36:00.1491358Z * [new branch] gh/soulitzer/349/base -> origin/gh/soulitzer/349/base 2025-09-07T07:36:00.1491907Z * [new branch] gh/soulitzer/349/head -> origin/gh/soulitzer/349/head 2025-09-07T07:36:00.1492573Z * [new branch] gh/soulitzer/349/orig -> origin/gh/soulitzer/349/orig 2025-09-07T07:36:00.1493363Z * [new branch] gh/soulitzer/350/base -> origin/gh/soulitzer/350/base 2025-09-07T07:36:00.1493772Z * [new branch] gh/soulitzer/350/head -> origin/gh/soulitzer/350/head 2025-09-07T07:36:00.1494371Z * [new branch] gh/soulitzer/350/orig -> origin/gh/soulitzer/350/orig 2025-09-07T07:36:00.1495137Z * [new branch] gh/soulitzer/351/base -> origin/gh/soulitzer/351/base 2025-09-07T07:36:00.1495682Z * [new branch] gh/soulitzer/351/head -> origin/gh/soulitzer/351/head 2025-09-07T07:36:00.1496101Z * [new branch] gh/soulitzer/351/orig -> origin/gh/soulitzer/351/orig 2025-09-07T07:36:00.1496901Z * [new branch] gh/soulitzer/353/base -> origin/gh/soulitzer/353/base 2025-09-07T07:36:00.1497483Z * [new branch] gh/soulitzer/353/head -> origin/gh/soulitzer/353/head 2025-09-07T07:36:00.1497953Z * [new branch] gh/soulitzer/353/orig -> origin/gh/soulitzer/353/orig 2025-09-07T07:36:00.1499358Z * [new branch] gh/soulitzer/358/base -> origin/gh/soulitzer/358/base 2025-09-07T07:36:00.1499908Z * [new branch] gh/soulitzer/358/head -> origin/gh/soulitzer/358/head 2025-09-07T07:36:00.1500492Z * [new branch] gh/soulitzer/358/orig -> origin/gh/soulitzer/358/orig 2025-09-07T07:36:00.1501644Z * [new branch] gh/soulitzer/359/base -> origin/gh/soulitzer/359/base 2025-09-07T07:36:00.1502294Z * [new branch] gh/soulitzer/359/head -> origin/gh/soulitzer/359/head 2025-09-07T07:36:00.1502835Z * [new branch] gh/soulitzer/359/orig -> origin/gh/soulitzer/359/orig 2025-09-07T07:36:00.1503613Z * [new branch] gh/soulitzer/362/base -> origin/gh/soulitzer/362/base 2025-09-07T07:36:00.1504180Z * [new branch] gh/soulitzer/362/head -> origin/gh/soulitzer/362/head 2025-09-07T07:36:00.1504598Z * [new branch] gh/soulitzer/362/orig -> origin/gh/soulitzer/362/orig 2025-09-07T07:36:00.1505408Z * [new branch] gh/soulitzer/372/base -> origin/gh/soulitzer/372/base 2025-09-07T07:36:00.1505815Z * [new branch] gh/soulitzer/372/head -> origin/gh/soulitzer/372/head 2025-09-07T07:36:00.1506436Z * [new branch] gh/soulitzer/372/orig -> origin/gh/soulitzer/372/orig 2025-09-07T07:36:00.1507195Z * [new branch] gh/soulitzer/373/base -> origin/gh/soulitzer/373/base 2025-09-07T07:36:00.1507749Z * [new branch] gh/soulitzer/373/head -> origin/gh/soulitzer/373/head 2025-09-07T07:36:00.1508206Z * [new branch] gh/soulitzer/373/orig -> origin/gh/soulitzer/373/orig 2025-09-07T07:36:00.1509084Z * [new branch] gh/soulitzer/374/base -> origin/gh/soulitzer/374/base 2025-09-07T07:36:00.1509528Z * [new branch] gh/soulitzer/374/head -> origin/gh/soulitzer/374/head 2025-09-07T07:36:00.1510138Z * [new branch] gh/soulitzer/374/orig -> origin/gh/soulitzer/374/orig 2025-09-07T07:36:00.1510877Z * [new branch] gh/soulitzer/375/base -> origin/gh/soulitzer/375/base 2025-09-07T07:36:00.1511519Z * [new branch] gh/soulitzer/375/head -> origin/gh/soulitzer/375/head 2025-09-07T07:36:00.1511971Z * [new branch] gh/soulitzer/375/orig -> origin/gh/soulitzer/375/orig 2025-09-07T07:36:00.1512813Z * [new branch] gh/soulitzer/376/base -> origin/gh/soulitzer/376/base 2025-09-07T07:36:00.1513259Z * [new branch] gh/soulitzer/376/head -> origin/gh/soulitzer/376/head 2025-09-07T07:36:00.1513901Z * [new branch] gh/soulitzer/376/orig -> origin/gh/soulitzer/376/orig 2025-09-07T07:36:00.1514620Z * [new branch] gh/soulitzer/377/base -> origin/gh/soulitzer/377/base 2025-09-07T07:36:00.1515178Z * [new branch] gh/soulitzer/377/head -> origin/gh/soulitzer/377/head 2025-09-07T07:36:00.1515573Z * [new branch] gh/soulitzer/377/orig -> origin/gh/soulitzer/377/orig 2025-09-07T07:36:00.1516659Z * [new branch] gh/soulitzer/378/base -> origin/gh/soulitzer/378/base 2025-09-07T07:36:00.1517035Z * [new branch] gh/soulitzer/378/head -> origin/gh/soulitzer/378/head 2025-09-07T07:36:00.1517648Z * [new branch] gh/soulitzer/378/orig -> origin/gh/soulitzer/378/orig 2025-09-07T07:36:00.1518375Z * [new branch] gh/soulitzer/379/base -> origin/gh/soulitzer/379/base 2025-09-07T07:36:00.1518815Z * [new branch] gh/soulitzer/379/head -> origin/gh/soulitzer/379/head 2025-09-07T07:36:00.1519406Z * [new branch] gh/soulitzer/379/orig -> origin/gh/soulitzer/379/orig 2025-09-07T07:36:00.1520351Z * [new branch] gh/swolchok/728/next -> origin/gh/swolchok/728/next 2025-09-07T07:36:00.1521475Z * [new branch] gh/swolchok/767/base -> origin/gh/swolchok/767/base 2025-09-07T07:36:00.1522172Z * [new branch] gh/swolchok/767/head -> origin/gh/swolchok/767/head 2025-09-07T07:36:00.1522968Z * [new branch] gh/swolchok/767/orig -> origin/gh/swolchok/767/orig 2025-09-07T07:36:00.1523724Z * [new branch] gh/swolchok/768/base -> origin/gh/swolchok/768/base 2025-09-07T07:36:00.1524383Z * [new branch] gh/swolchok/768/head -> origin/gh/swolchok/768/head 2025-09-07T07:36:00.1524980Z * [new branch] gh/swolchok/768/orig -> origin/gh/swolchok/768/orig 2025-09-07T07:36:00.1525866Z * [new branch] gh/swolchok/769/base -> origin/gh/swolchok/769/base 2025-09-07T07:36:00.1526420Z * [new branch] gh/swolchok/769/head -> origin/gh/swolchok/769/head 2025-09-07T07:36:00.1527061Z * [new branch] gh/swolchok/769/orig -> origin/gh/swolchok/769/orig 2025-09-07T07:36:00.1527868Z * [new branch] gh/swolchok/771/base -> origin/gh/swolchok/771/base 2025-09-07T07:36:00.1528433Z * [new branch] gh/swolchok/771/head -> origin/gh/swolchok/771/head 2025-09-07T07:36:00.1528871Z * [new branch] gh/swolchok/771/orig -> origin/gh/swolchok/771/orig 2025-09-07T07:36:00.1529694Z * [new branch] gh/swolchok/772/base -> origin/gh/swolchok/772/base 2025-09-07T07:36:00.1530343Z * [new branch] gh/swolchok/772/head -> origin/gh/swolchok/772/head 2025-09-07T07:36:00.1530914Z * [new branch] gh/swolchok/772/orig -> origin/gh/swolchok/772/orig 2025-09-07T07:36:00.1531844Z * [new branch] gh/swolchok/773/base -> origin/gh/swolchok/773/base 2025-09-07T07:36:00.1532460Z * [new branch] gh/swolchok/773/head -> origin/gh/swolchok/773/head 2025-09-07T07:36:00.1533060Z * [new branch] gh/swolchok/773/orig -> origin/gh/swolchok/773/orig 2025-09-07T07:36:00.1533804Z * [new branch] gh/swolchok/786/base -> origin/gh/swolchok/786/base 2025-09-07T07:36:00.1534302Z * [new branch] gh/swolchok/786/head -> origin/gh/swolchok/786/head 2025-09-07T07:36:00.1534918Z * [new branch] gh/swolchok/786/orig -> origin/gh/swolchok/786/orig 2025-09-07T07:36:00.1535940Z * [new branch] gh/swolchok/787/base -> origin/gh/swolchok/787/base 2025-09-07T07:36:00.1536483Z * [new branch] gh/swolchok/787/head -> origin/gh/swolchok/787/head 2025-09-07T07:36:00.1537029Z * [new branch] gh/swolchok/787/orig -> origin/gh/swolchok/787/orig 2025-09-07T07:36:00.1537786Z * [new branch] gh/swolchok/788/base -> origin/gh/swolchok/788/base 2025-09-07T07:36:00.1538379Z * [new branch] gh/swolchok/788/head -> origin/gh/swolchok/788/head 2025-09-07T07:36:00.1538765Z * [new branch] gh/swolchok/788/orig -> origin/gh/swolchok/788/orig 2025-09-07T07:36:00.1539545Z * [new branch] gh/swolchok/789/base -> origin/gh/swolchok/789/base 2025-09-07T07:36:00.1540147Z * [new branch] gh/swolchok/789/head -> origin/gh/swolchok/789/head 2025-09-07T07:36:00.1540726Z * [new branch] gh/swolchok/789/orig -> origin/gh/swolchok/789/orig 2025-09-07T07:36:00.1541568Z * [new branch] gh/swolchok/790/base -> origin/gh/swolchok/790/base 2025-09-07T07:36:00.1542117Z * [new branch] gh/swolchok/790/head -> origin/gh/swolchok/790/head 2025-09-07T07:36:00.1542659Z * [new branch] gh/swolchok/790/orig -> origin/gh/swolchok/790/orig 2025-09-07T07:36:00.1543787Z * [new branch] gh/swolchok/791/base -> origin/gh/swolchok/791/base 2025-09-07T07:36:00.1544170Z * [new branch] gh/swolchok/791/head -> origin/gh/swolchok/791/head 2025-09-07T07:36:00.1544859Z * [new branch] gh/swolchok/791/orig -> origin/gh/swolchok/791/orig 2025-09-07T07:36:00.1545935Z * [new branch] gh/swolchok/792/base -> origin/gh/swolchok/792/base 2025-09-07T07:36:00.1546404Z * [new branch] gh/swolchok/792/head -> origin/gh/swolchok/792/head 2025-09-07T07:36:00.1547006Z * [new branch] gh/swolchok/792/orig -> origin/gh/swolchok/792/orig 2025-09-07T07:36:00.1547742Z * [new branch] gh/swolchok/793/base -> origin/gh/swolchok/793/base 2025-09-07T07:36:00.1548345Z * [new branch] gh/swolchok/793/head -> origin/gh/swolchok/793/head 2025-09-07T07:36:00.1548747Z * [new branch] gh/swolchok/793/orig -> origin/gh/swolchok/793/orig 2025-09-07T07:36:00.1549628Z * [new branch] gh/swolchok/794/base -> origin/gh/swolchok/794/base 2025-09-07T07:36:00.1550071Z * [new branch] gh/swolchok/794/head -> origin/gh/swolchok/794/head 2025-09-07T07:36:00.1550766Z * [new branch] gh/swolchok/794/orig -> origin/gh/swolchok/794/orig 2025-09-07T07:36:00.1551840Z * [new branch] gh/swolchok/795/base -> origin/gh/swolchok/795/base 2025-09-07T07:36:00.1552385Z * [new branch] gh/swolchok/795/head -> origin/gh/swolchok/795/head 2025-09-07T07:36:00.1552992Z * [new branch] gh/swolchok/795/orig -> origin/gh/swolchok/795/orig 2025-09-07T07:36:00.1553817Z * [new branch] gh/swolchok/796/base -> origin/gh/swolchok/796/base 2025-09-07T07:36:00.1554359Z * [new branch] gh/swolchok/796/head -> origin/gh/swolchok/796/head 2025-09-07T07:36:00.1554964Z * [new branch] gh/swolchok/796/orig -> origin/gh/swolchok/796/orig 2025-09-07T07:36:00.1555749Z * [new branch] gh/swolchok/797/base -> origin/gh/swolchok/797/base 2025-09-07T07:36:00.1556454Z * [new branch] gh/swolchok/797/head -> origin/gh/swolchok/797/head 2025-09-07T07:36:00.1556878Z * [new branch] gh/swolchok/797/orig -> origin/gh/swolchok/797/orig 2025-09-07T07:36:00.1557759Z * [new branch] gh/swolchok/798/base -> origin/gh/swolchok/798/base 2025-09-07T07:36:00.1558190Z * [new branch] gh/swolchok/798/head -> origin/gh/swolchok/798/head 2025-09-07T07:36:00.1558827Z * [new branch] gh/swolchok/798/orig -> origin/gh/swolchok/798/orig 2025-09-07T07:36:00.1560016Z * [new branch] gh/swolchok/799/base -> origin/gh/swolchok/799/base 2025-09-07T07:36:00.1560452Z * [new branch] gh/swolchok/799/head -> origin/gh/swolchok/799/head 2025-09-07T07:36:00.1561272Z * [new branch] gh/swolchok/799/orig -> origin/gh/swolchok/799/orig 2025-09-07T07:36:00.1562153Z * [new branch] gh/swolchok/800/base -> origin/gh/swolchok/800/base 2025-09-07T07:36:00.1562599Z * [new branch] gh/swolchok/800/head -> origin/gh/swolchok/800/head 2025-09-07T07:36:00.1563244Z * [new branch] gh/swolchok/800/orig -> origin/gh/swolchok/800/orig 2025-09-07T07:36:00.1564071Z * [new branch] gh/swolchok/801/base -> origin/gh/swolchok/801/base 2025-09-07T07:36:00.1564614Z * [new branch] gh/swolchok/801/head -> origin/gh/swolchok/801/head 2025-09-07T07:36:00.1565212Z * [new branch] gh/swolchok/801/orig -> origin/gh/swolchok/801/orig 2025-09-07T07:36:00.1565970Z * [new branch] gh/swolchok/802/base -> origin/gh/swolchok/802/base 2025-09-07T07:36:00.1566357Z * [new branch] gh/swolchok/802/head -> origin/gh/swolchok/802/head 2025-09-07T07:36:00.1566996Z * [new branch] gh/swolchok/802/orig -> origin/gh/swolchok/802/orig 2025-09-07T07:36:00.1567726Z * [new branch] gh/swolchok/803/base -> origin/gh/swolchok/803/base 2025-09-07T07:36:00.1568169Z * [new branch] gh/swolchok/803/head -> origin/gh/swolchok/803/head 2025-09-07T07:36:00.1568864Z * [new branch] gh/swolchok/803/orig -> origin/gh/swolchok/803/orig 2025-09-07T07:36:00.1569792Z * [new branch] gh/swolchok/804/base -> origin/gh/swolchok/804/base 2025-09-07T07:36:00.1570150Z * [new branch] gh/swolchok/804/head -> origin/gh/swolchok/804/head 2025-09-07T07:36:00.1571000Z * [new branch] gh/swolchok/804/orig -> origin/gh/swolchok/804/orig 2025-09-07T07:36:00.1571763Z * [new branch] gh/swolchok/805/base -> origin/gh/swolchok/805/base 2025-09-07T07:36:00.1572366Z * [new branch] gh/swolchok/805/head -> origin/gh/swolchok/805/head 2025-09-07T07:36:00.1572905Z * [new branch] gh/swolchok/805/orig -> origin/gh/swolchok/805/orig 2025-09-07T07:36:00.1573564Z * [new branch] gh/swolchok/806/base -> origin/gh/swolchok/806/base 2025-09-07T07:36:00.1574123Z * [new branch] gh/swolchok/806/head -> origin/gh/swolchok/806/head 2025-09-07T07:36:00.1574560Z * [new branch] gh/swolchok/806/orig -> origin/gh/swolchok/806/orig 2025-09-07T07:36:00.1575524Z * [new branch] gh/swolchok/807/base -> origin/gh/swolchok/807/base 2025-09-07T07:36:00.1575953Z * [new branch] gh/swolchok/807/head -> origin/gh/swolchok/807/head 2025-09-07T07:36:00.1576598Z * [new branch] gh/swolchok/807/orig -> origin/gh/swolchok/807/orig 2025-09-07T07:36:00.1577793Z * [new branch] gh/swolchok/808/base -> origin/gh/swolchok/808/base 2025-09-07T07:36:00.1578337Z * [new branch] gh/swolchok/808/head -> origin/gh/swolchok/808/head 2025-09-07T07:36:00.1578772Z * [new branch] gh/swolchok/808/orig -> origin/gh/swolchok/808/orig 2025-09-07T07:36:00.1579603Z * [new branch] gh/swolchok/809/base -> origin/gh/swolchok/809/base 2025-09-07T07:36:00.1580219Z * [new branch] gh/swolchok/809/head -> origin/gh/swolchok/809/head 2025-09-07T07:36:00.1580898Z * [new branch] gh/swolchok/809/orig -> origin/gh/swolchok/809/orig 2025-09-07T07:36:00.1581739Z * [new branch] gh/swolchok/810/base -> origin/gh/swolchok/810/base 2025-09-07T07:36:00.1582157Z * [new branch] gh/swolchok/810/head -> origin/gh/swolchok/810/head 2025-09-07T07:36:00.1582745Z * [new branch] gh/swolchok/810/orig -> origin/gh/swolchok/810/orig 2025-09-07T07:36:00.1583543Z * [new branch] gh/swolchok/811/base -> origin/gh/swolchok/811/base 2025-09-07T07:36:00.1584149Z * [new branch] gh/swolchok/811/head -> origin/gh/swolchok/811/head 2025-09-07T07:36:00.1584731Z * [new branch] gh/swolchok/811/orig -> origin/gh/swolchok/811/orig 2025-09-07T07:36:00.1585533Z * [new branch] gh/swolchok/812/base -> origin/gh/swolchok/812/base 2025-09-07T07:36:00.1585922Z * [new branch] gh/swolchok/812/head -> origin/gh/swolchok/812/head 2025-09-07T07:36:00.1586520Z * [new branch] gh/swolchok/812/orig -> origin/gh/swolchok/812/orig 2025-09-07T07:36:00.1587298Z * [new branch] gh/swolchok/813/base -> origin/gh/swolchok/813/base 2025-09-07T07:36:00.1587843Z * [new branch] gh/swolchok/813/head -> origin/gh/swolchok/813/head 2025-09-07T07:36:00.1588538Z * [new branch] gh/swolchok/813/orig -> origin/gh/swolchok/813/orig 2025-09-07T07:36:00.1589887Z * [new branch] gh/swolchok/814/base -> origin/gh/swolchok/814/base 2025-09-07T07:36:00.1590446Z * [new branch] gh/swolchok/814/head -> origin/gh/swolchok/814/head 2025-09-07T07:36:00.1591050Z * [new branch] gh/swolchok/814/orig -> origin/gh/swolchok/814/orig 2025-09-07T07:36:00.1591915Z * [new branch] gh/swolchok/815/base -> origin/gh/swolchok/815/base 2025-09-07T07:36:00.1592302Z * [new branch] gh/swolchok/815/head -> origin/gh/swolchok/815/head 2025-09-07T07:36:00.1592963Z * [new branch] gh/swolchok/815/orig -> origin/gh/swolchok/815/orig 2025-09-07T07:36:00.1593740Z * [new branch] gh/swolchok/816/base -> origin/gh/swolchok/816/base 2025-09-07T07:36:00.1594356Z * [new branch] gh/swolchok/816/head -> origin/gh/swolchok/816/head 2025-09-07T07:36:00.1594947Z * [new branch] gh/swolchok/816/orig -> origin/gh/swolchok/816/orig 2025-09-07T07:36:00.1595757Z * [new branch] gh/swolchok/817/base -> origin/gh/swolchok/817/base 2025-09-07T07:36:00.1596292Z * [new branch] gh/swolchok/817/head -> origin/gh/swolchok/817/head 2025-09-07T07:36:00.1596735Z * [new branch] gh/swolchok/817/orig -> origin/gh/swolchok/817/orig 2025-09-07T07:36:00.1597586Z * [new branch] gh/swolchok/818/base -> origin/gh/swolchok/818/base 2025-09-07T07:36:00.1598021Z * [new branch] gh/swolchok/818/head -> origin/gh/swolchok/818/head 2025-09-07T07:36:00.1598616Z * [new branch] gh/swolchok/818/orig -> origin/gh/swolchok/818/orig 2025-09-07T07:36:00.1599476Z * [new branch] gh/swolchok/819/base -> origin/gh/swolchok/819/base 2025-09-07T07:36:00.1600072Z * [new branch] gh/swolchok/819/head -> origin/gh/swolchok/819/head 2025-09-07T07:36:00.1600635Z * [new branch] gh/swolchok/819/orig -> origin/gh/swolchok/819/orig 2025-09-07T07:36:00.1601385Z * [new branch] gh/swolchok/820/base -> origin/gh/swolchok/820/base 2025-09-07T07:36:00.1601764Z * [new branch] gh/swolchok/820/head -> origin/gh/swolchok/820/head 2025-09-07T07:36:00.1602375Z * [new branch] gh/swolchok/820/orig -> origin/gh/swolchok/820/orig 2025-09-07T07:36:00.1603150Z * [new branch] gh/swolchok/821/base -> origin/gh/swolchok/821/base 2025-09-07T07:36:00.1603744Z * [new branch] gh/swolchok/821/head -> origin/gh/swolchok/821/head 2025-09-07T07:36:00.1604161Z * [new branch] gh/swolchok/821/orig -> origin/gh/swolchok/821/orig 2025-09-07T07:36:00.1605149Z * [new branch] gh/swolchok/822/base -> origin/gh/swolchok/822/base 2025-09-07T07:36:00.1605576Z * [new branch] gh/swolchok/822/head -> origin/gh/swolchok/822/head 2025-09-07T07:36:00.1606270Z * [new branch] gh/swolchok/822/orig -> origin/gh/swolchok/822/orig 2025-09-07T07:36:00.1606970Z * [new branch] gh/swolchok/823/base -> origin/gh/swolchok/823/base 2025-09-07T07:36:00.1607351Z * [new branch] gh/swolchok/823/head -> origin/gh/swolchok/823/head 2025-09-07T07:36:00.1607958Z * [new branch] gh/swolchok/823/orig -> origin/gh/swolchok/823/orig 2025-09-07T07:36:00.1608999Z * [new branch] gh/swolchok/824/base -> origin/gh/swolchok/824/base 2025-09-07T07:36:00.1609661Z * [new branch] gh/swolchok/824/head -> origin/gh/swolchok/824/head 2025-09-07T07:36:00.1610216Z * [new branch] gh/swolchok/824/orig -> origin/gh/swolchok/824/orig 2025-09-07T07:36:00.1610930Z * [new branch] gh/swolchok/825/base -> origin/gh/swolchok/825/base 2025-09-07T07:36:00.1611544Z * [new branch] gh/swolchok/825/head -> origin/gh/swolchok/825/head 2025-09-07T07:36:00.1612127Z * [new branch] gh/swolchok/825/orig -> origin/gh/swolchok/825/orig 2025-09-07T07:36:00.1612945Z * [new branch] gh/swolchok/826/base -> origin/gh/swolchok/826/base 2025-09-07T07:36:00.1613343Z * [new branch] gh/swolchok/826/head -> origin/gh/swolchok/826/head 2025-09-07T07:36:00.1613895Z * [new branch] gh/swolchok/826/orig -> origin/gh/swolchok/826/orig 2025-09-07T07:36:00.1614711Z * [new branch] gh/swolchok/827/base -> origin/gh/swolchok/827/base 2025-09-07T07:36:00.1615177Z * [new branch] gh/swolchok/827/head -> origin/gh/swolchok/827/head 2025-09-07T07:36:00.1615585Z * [new branch] gh/swolchok/827/orig -> origin/gh/swolchok/827/orig 2025-09-07T07:36:00.1616493Z * [new branch] gh/swolchok/828/base -> origin/gh/swolchok/828/base 2025-09-07T07:36:00.1616871Z * [new branch] gh/swolchok/828/head -> origin/gh/swolchok/828/head 2025-09-07T07:36:00.1617453Z * [new branch] gh/swolchok/828/orig -> origin/gh/swolchok/828/orig 2025-09-07T07:36:00.1618094Z * [new branch] gh/swolchok/829/base -> origin/gh/swolchok/829/base 2025-09-07T07:36:00.1618801Z * [new branch] gh/swolchok/829/head -> origin/gh/swolchok/829/head 2025-09-07T07:36:00.1619346Z * [new branch] gh/swolchok/829/orig -> origin/gh/swolchok/829/orig 2025-09-07T07:36:00.1620197Z * [new branch] gh/swolchok/830/base -> origin/gh/swolchok/830/base 2025-09-07T07:36:00.1620603Z * [new branch] gh/swolchok/830/head -> origin/gh/swolchok/830/head 2025-09-07T07:36:00.1621146Z * [new branch] gh/swolchok/830/orig -> origin/gh/swolchok/830/orig 2025-09-07T07:36:00.1622095Z * [new branch] gh/swolchok/831/base -> origin/gh/swolchok/831/base 2025-09-07T07:36:00.1622695Z * [new branch] gh/swolchok/831/head -> origin/gh/swolchok/831/head 2025-09-07T07:36:00.1623118Z * [new branch] gh/swolchok/831/orig -> origin/gh/swolchok/831/orig 2025-09-07T07:36:00.1624144Z * [new branch] gh/swolchok/832/base -> origin/gh/swolchok/832/base 2025-09-07T07:36:00.1624771Z * [new branch] gh/swolchok/832/head -> origin/gh/swolchok/832/head 2025-09-07T07:36:00.1625151Z * [new branch] gh/swolchok/832/orig -> origin/gh/swolchok/832/orig 2025-09-07T07:36:00.1626257Z * [new branch] gh/syed-ahmed/3/base -> origin/gh/syed-ahmed/3/base 2025-09-07T07:36:00.1626691Z * [new branch] gh/syed-ahmed/3/head -> origin/gh/syed-ahmed/3/head 2025-09-07T07:36:00.1627653Z * [new branch] gh/syed-ahmed/3/orig -> origin/gh/syed-ahmed/3/orig 2025-09-07T07:36:00.1628383Z * [new branch] gh/syed-ahmed/4/base -> origin/gh/syed-ahmed/4/base 2025-09-07T07:36:00.1628987Z * [new branch] gh/syed-ahmed/4/head -> origin/gh/syed-ahmed/4/head 2025-09-07T07:36:00.1629533Z * [new branch] gh/syed-ahmed/4/orig -> origin/gh/syed-ahmed/4/orig 2025-09-07T07:36:00.1630261Z * [new branch] gh/syed-ahmed/5/base -> origin/gh/syed-ahmed/5/base 2025-09-07T07:36:00.1630707Z * [new branch] gh/syed-ahmed/5/head -> origin/gh/syed-ahmed/5/head 2025-09-07T07:36:00.1631294Z * [new branch] gh/syed-ahmed/5/orig -> origin/gh/syed-ahmed/5/orig 2025-09-07T07:36:00.1632302Z * [new branch] gh/teja-rao/4/base -> origin/gh/teja-rao/4/base 2025-09-07T07:36:00.1632849Z * [new branch] gh/teja-rao/4/head -> origin/gh/teja-rao/4/head 2025-09-07T07:36:00.1633448Z * [new branch] gh/teja-rao/4/orig -> origin/gh/teja-rao/4/orig 2025-09-07T07:36:00.1634711Z * [new branch] gh/tianyu-l/2/base -> origin/gh/tianyu-l/2/base 2025-09-07T07:36:00.1635340Z * [new branch] gh/tianyu-l/2/head -> origin/gh/tianyu-l/2/head 2025-09-07T07:36:00.1635704Z * [new branch] gh/tianyu-l/2/orig -> origin/gh/tianyu-l/2/orig 2025-09-07T07:36:00.1636480Z * [new branch] gh/tianyu-l/3/base -> origin/gh/tianyu-l/3/base 2025-09-07T07:36:00.1637030Z * [new branch] gh/tianyu-l/3/head -> origin/gh/tianyu-l/3/head 2025-09-07T07:36:00.1637466Z * [new branch] gh/tianyu-l/3/orig -> origin/gh/tianyu-l/3/orig 2025-09-07T07:36:00.1638340Z * [new branch] gh/tianyu-l/4/base -> origin/gh/tianyu-l/4/base 2025-09-07T07:36:00.1638935Z * [new branch] gh/tianyu-l/4/head -> origin/gh/tianyu-l/4/head 2025-09-07T07:36:00.1639516Z * [new branch] gh/tianyu-l/4/orig -> origin/gh/tianyu-l/4/orig 2025-09-07T07:36:00.1640570Z * [new branch] gh/tugsbayasgalan/1/base -> origin/gh/tugsbayasgalan/1/base 2025-09-07T07:36:00.1640996Z * [new branch] gh/tugsbayasgalan/1/head -> origin/gh/tugsbayasgalan/1/head 2025-09-07T07:36:00.1641674Z * [new branch] gh/tugsbayasgalan/1/orig -> origin/gh/tugsbayasgalan/1/orig 2025-09-07T07:36:00.1642706Z * [new branch] gh/tugsbayasgalan/10/base -> origin/gh/tugsbayasgalan/10/base 2025-09-07T07:36:00.1643296Z * [new branch] gh/tugsbayasgalan/10/head -> origin/gh/tugsbayasgalan/10/head 2025-09-07T07:36:00.1643694Z * [new branch] gh/tugsbayasgalan/10/orig -> origin/gh/tugsbayasgalan/10/orig 2025-09-07T07:36:00.1644453Z * [new branch] gh/tugsbayasgalan/11/base -> origin/gh/tugsbayasgalan/11/base 2025-09-07T07:36:00.1645272Z * [new branch] gh/tugsbayasgalan/11/head -> origin/gh/tugsbayasgalan/11/head 2025-09-07T07:36:00.1645935Z * [new branch] gh/tugsbayasgalan/11/orig -> origin/gh/tugsbayasgalan/11/orig 2025-09-07T07:36:00.1646619Z * [new branch] gh/tugsbayasgalan/12/base -> origin/gh/tugsbayasgalan/12/base 2025-09-07T07:36:00.1647031Z * [new branch] gh/tugsbayasgalan/12/head -> origin/gh/tugsbayasgalan/12/head 2025-09-07T07:36:00.1647650Z * [new branch] gh/tugsbayasgalan/12/orig -> origin/gh/tugsbayasgalan/12/orig 2025-09-07T07:36:00.1648453Z * [new branch] gh/tugsbayasgalan/13/base -> origin/gh/tugsbayasgalan/13/base 2025-09-07T07:36:00.1649001Z * [new branch] gh/tugsbayasgalan/13/head -> origin/gh/tugsbayasgalan/13/head 2025-09-07T07:36:00.1649622Z * [new branch] gh/tugsbayasgalan/13/orig -> origin/gh/tugsbayasgalan/13/orig 2025-09-07T07:36:00.1650478Z * [new branch] gh/tugsbayasgalan/14/base -> origin/gh/tugsbayasgalan/14/base 2025-09-07T07:36:00.1650933Z * [new branch] gh/tugsbayasgalan/14/head -> origin/gh/tugsbayasgalan/14/head 2025-09-07T07:36:00.1651498Z * [new branch] gh/tugsbayasgalan/14/orig -> origin/gh/tugsbayasgalan/14/orig 2025-09-07T07:36:00.1652634Z * [new branch] gh/tugsbayasgalan/15/base -> origin/gh/tugsbayasgalan/15/base 2025-09-07T07:36:00.1653044Z * [new branch] gh/tugsbayasgalan/15/head -> origin/gh/tugsbayasgalan/15/head 2025-09-07T07:36:00.1653601Z * [new branch] gh/tugsbayasgalan/15/orig -> origin/gh/tugsbayasgalan/15/orig 2025-09-07T07:36:00.1654436Z * [new branch] gh/tugsbayasgalan/2/base -> origin/gh/tugsbayasgalan/2/base 2025-09-07T07:36:00.1654857Z * [new branch] gh/tugsbayasgalan/2/head -> origin/gh/tugsbayasgalan/2/head 2025-09-07T07:36:00.1655421Z * [new branch] gh/tugsbayasgalan/2/orig -> origin/gh/tugsbayasgalan/2/orig 2025-09-07T07:36:00.1656082Z * [new branch] gh/tugsbayasgalan/3/base -> origin/gh/tugsbayasgalan/3/base 2025-09-07T07:36:00.1656888Z * [new branch] gh/tugsbayasgalan/3/head -> origin/gh/tugsbayasgalan/3/head 2025-09-07T07:36:00.1657294Z * [new branch] gh/tugsbayasgalan/3/orig -> origin/gh/tugsbayasgalan/3/orig 2025-09-07T07:36:00.1658209Z * [new branch] gh/tugsbayasgalan/4/base -> origin/gh/tugsbayasgalan/4/base 2025-09-07T07:36:00.1658961Z * [new branch] gh/tugsbayasgalan/4/head -> origin/gh/tugsbayasgalan/4/head 2025-09-07T07:36:00.1659357Z * [new branch] gh/tugsbayasgalan/4/orig -> origin/gh/tugsbayasgalan/4/orig 2025-09-07T07:36:00.1660349Z * [new branch] gh/tugsbayasgalan/5/base -> origin/gh/tugsbayasgalan/5/base 2025-09-07T07:36:00.1660820Z * [new branch] gh/tugsbayasgalan/5/head -> origin/gh/tugsbayasgalan/5/head 2025-09-07T07:36:00.1661387Z * [new branch] gh/tugsbayasgalan/5/orig -> origin/gh/tugsbayasgalan/5/orig 2025-09-07T07:36:00.1662067Z * [new branch] gh/tugsbayasgalan/6/base -> origin/gh/tugsbayasgalan/6/base 2025-09-07T07:36:00.1662462Z * [new branch] gh/tugsbayasgalan/6/head -> origin/gh/tugsbayasgalan/6/head 2025-09-07T07:36:00.1663086Z * [new branch] gh/tugsbayasgalan/6/orig -> origin/gh/tugsbayasgalan/6/orig 2025-09-07T07:36:00.1663922Z * [new branch] gh/tugsbayasgalan/7/base -> origin/gh/tugsbayasgalan/7/base 2025-09-07T07:36:00.1664536Z * [new branch] gh/tugsbayasgalan/7/head -> origin/gh/tugsbayasgalan/7/head 2025-09-07T07:36:00.1664947Z * [new branch] gh/tugsbayasgalan/7/orig -> origin/gh/tugsbayasgalan/7/orig 2025-09-07T07:36:00.1665728Z * [new branch] gh/tugsbayasgalan/8/base -> origin/gh/tugsbayasgalan/8/base 2025-09-07T07:36:00.1666150Z * [new branch] gh/tugsbayasgalan/8/head -> origin/gh/tugsbayasgalan/8/head 2025-09-07T07:36:00.1666711Z * [new branch] gh/tugsbayasgalan/8/orig -> origin/gh/tugsbayasgalan/8/orig 2025-09-07T07:36:00.1667576Z * [new branch] gh/tugsbayasgalan/9/base -> origin/gh/tugsbayasgalan/9/base 2025-09-07T07:36:00.1667929Z * [new branch] gh/tugsbayasgalan/9/head -> origin/gh/tugsbayasgalan/9/head 2025-09-07T07:36:00.1668589Z * [new branch] gh/tugsbayasgalan/9/orig -> origin/gh/tugsbayasgalan/9/orig 2025-09-07T07:36:00.1673545Z * [new branch] gh/v0i0/1/base -> origin/gh/v0i0/1/base 2025-09-07T07:36:00.1674086Z * [new branch] gh/v0i0/1/head -> origin/gh/v0i0/1/head 2025-09-07T07:36:00.1674548Z * [new branch] gh/v0i0/1/orig -> origin/gh/v0i0/1/orig 2025-09-07T07:36:00.1675521Z * [new branch] gh/v0i0/4/base -> origin/gh/v0i0/4/base 2025-09-07T07:36:00.1675905Z * [new branch] gh/v0i0/4/head -> origin/gh/v0i0/4/head 2025-09-07T07:36:00.1676478Z * [new branch] gh/v0i0/4/orig -> origin/gh/v0i0/4/orig 2025-09-07T07:36:00.1677183Z * [new branch] gh/v0i0/6/base -> origin/gh/v0i0/6/base 2025-09-07T07:36:00.1677720Z * [new branch] gh/v0i0/6/head -> origin/gh/v0i0/6/head 2025-09-07T07:36:00.1678331Z * [new branch] gh/v0i0/6/orig -> origin/gh/v0i0/6/orig 2025-09-07T07:36:00.1679059Z * [new branch] gh/v0i0/7/base -> origin/gh/v0i0/7/base 2025-09-07T07:36:00.1679652Z * [new branch] gh/v0i0/7/head -> origin/gh/v0i0/7/head 2025-09-07T07:36:00.1680034Z * [new branch] gh/v0i0/7/orig -> origin/gh/v0i0/7/orig 2025-09-07T07:36:00.1680918Z * [new branch] gh/v0i0/8/base -> origin/gh/v0i0/8/base 2025-09-07T07:36:00.1681453Z * [new branch] gh/v0i0/8/head -> origin/gh/v0i0/8/head 2025-09-07T07:36:00.1681872Z * [new branch] gh/v0i0/8/orig -> origin/gh/v0i0/8/orig 2025-09-07T07:36:00.1682827Z * [new branch] gh/v0i0/9/base -> origin/gh/v0i0/9/base 2025-09-07T07:36:00.1683213Z * [new branch] gh/v0i0/9/head -> origin/gh/v0i0/9/head 2025-09-07T07:36:00.1683810Z * [new branch] gh/v0i0/9/orig -> origin/gh/v0i0/9/orig 2025-09-07T07:36:00.1685107Z * [new branch] gh/vkuzo/1/next -> origin/gh/vkuzo/1/next 2025-09-07T07:36:00.1685841Z * [new branch] gh/vkuzo/2/next -> origin/gh/vkuzo/2/next 2025-09-07T07:36:00.1686509Z * [new branch] gh/vkuzo/3/next -> origin/gh/vkuzo/3/next 2025-09-07T07:36:00.1687370Z * [new branch] gh/vkuzo/4/base -> origin/gh/vkuzo/4/base 2025-09-07T07:36:00.1687960Z * [new branch] gh/vkuzo/4/head -> origin/gh/vkuzo/4/head 2025-09-07T07:36:00.1688499Z * [new branch] gh/vkuzo/4/orig -> origin/gh/vkuzo/4/orig 2025-09-07T07:36:00.1689372Z * [new branch] gh/vkuzo/5/base -> origin/gh/vkuzo/5/base 2025-09-07T07:36:00.1690102Z * [new branch] gh/vkuzo/5/head -> origin/gh/vkuzo/5/head 2025-09-07T07:36:00.1690576Z * [new branch] gh/vkuzo/5/orig -> origin/gh/vkuzo/5/orig 2025-09-07T07:36:00.1691619Z * [new branch] gh/vkuzo/6/base -> origin/gh/vkuzo/6/base 2025-09-07T07:36:00.1692002Z * [new branch] gh/vkuzo/6/head -> origin/gh/vkuzo/6/head 2025-09-07T07:36:00.1692639Z * [new branch] gh/vkuzo/6/orig -> origin/gh/vkuzo/6/orig 2025-09-07T07:36:00.1693307Z * [new branch] gh/vkuzo/7/base -> origin/gh/vkuzo/7/base 2025-09-07T07:36:00.1693941Z * [new branch] gh/vkuzo/7/head -> origin/gh/vkuzo/7/head 2025-09-07T07:36:00.1694486Z * [new branch] gh/vkuzo/7/orig -> origin/gh/vkuzo/7/orig 2025-09-07T07:36:00.1695449Z * [new branch] gh/wconstab/419/base -> origin/gh/wconstab/419/base 2025-09-07T07:36:00.1695874Z * [new branch] gh/wconstab/419/head -> origin/gh/wconstab/419/head 2025-09-07T07:36:00.1696855Z * [new branch] gh/wconstab/419/orig -> origin/gh/wconstab/419/orig 2025-09-07T07:36:00.1697711Z * [new branch] gh/wconstab/424/base -> origin/gh/wconstab/424/base 2025-09-07T07:36:00.1698098Z * [new branch] gh/wconstab/424/head -> origin/gh/wconstab/424/head 2025-09-07T07:36:00.1698712Z * [new branch] gh/wconstab/424/orig -> origin/gh/wconstab/424/orig 2025-09-07T07:36:00.1699448Z * [new branch] gh/wconstab/435/base -> origin/gh/wconstab/435/base 2025-09-07T07:36:00.1699994Z * [new branch] gh/wconstab/435/head -> origin/gh/wconstab/435/head 2025-09-07T07:36:00.1700682Z * [new branch] gh/wconstab/435/orig -> origin/gh/wconstab/435/orig 2025-09-07T07:36:00.1701447Z * [new branch] gh/wconstab/438/base -> origin/gh/wconstab/438/base 2025-09-07T07:36:00.1701892Z * [new branch] gh/wconstab/438/head -> origin/gh/wconstab/438/head 2025-09-07T07:36:00.1702484Z * [new branch] gh/wconstab/438/orig -> origin/gh/wconstab/438/orig 2025-09-07T07:36:00.1703232Z * [new branch] gh/wconstab/440/base -> origin/gh/wconstab/440/base 2025-09-07T07:36:00.1703887Z * [new branch] gh/wconstab/440/head -> origin/gh/wconstab/440/head 2025-09-07T07:36:00.1704629Z * [new branch] gh/wconstab/440/orig -> origin/gh/wconstab/440/orig 2025-09-07T07:36:00.1705834Z * [new branch] gh/wconstab/441/base -> origin/gh/wconstab/441/base 2025-09-07T07:36:00.1706165Z * [new branch] gh/wconstab/441/head -> origin/gh/wconstab/441/head 2025-09-07T07:36:00.1706876Z * [new branch] gh/wconstab/441/orig -> origin/gh/wconstab/441/orig 2025-09-07T07:36:00.1707683Z * [new branch] gh/wconstab/442/base -> origin/gh/wconstab/442/base 2025-09-07T07:36:00.1708280Z * [new branch] gh/wconstab/442/head -> origin/gh/wconstab/442/head 2025-09-07T07:36:00.1708876Z * [new branch] gh/wconstab/442/orig -> origin/gh/wconstab/442/orig 2025-09-07T07:36:00.1709640Z * [new branch] gh/wconstab/443/base -> origin/gh/wconstab/443/base 2025-09-07T07:36:00.1710190Z * [new branch] gh/wconstab/443/head -> origin/gh/wconstab/443/head 2025-09-07T07:36:00.1710790Z * [new branch] gh/wconstab/443/orig -> origin/gh/wconstab/443/orig 2025-09-07T07:36:00.1711558Z * [new branch] gh/wconstab/444/base -> origin/gh/wconstab/444/base 2025-09-07T07:36:00.1712098Z * [new branch] gh/wconstab/444/head -> origin/gh/wconstab/444/head 2025-09-07T07:36:00.1712654Z * [new branch] gh/wconstab/444/orig -> origin/gh/wconstab/444/orig 2025-09-07T07:36:00.1713448Z * [new branch] gh/wconstab/445/base -> origin/gh/wconstab/445/base 2025-09-07T07:36:00.1713889Z * [new branch] gh/wconstab/445/head -> origin/gh/wconstab/445/head 2025-09-07T07:36:00.1714482Z * [new branch] gh/wconstab/445/orig -> origin/gh/wconstab/445/orig 2025-09-07T07:36:00.1715631Z * [new branch] gh/wconstab/446/base -> origin/gh/wconstab/446/base 2025-09-07T07:36:00.1716337Z * [new branch] gh/wconstab/446/head -> origin/gh/wconstab/446/head 2025-09-07T07:36:00.1717158Z * [new branch] gh/wconstab/446/orig -> origin/gh/wconstab/446/orig 2025-09-07T07:36:00.1717919Z * [new branch] gh/wconstab/447/base -> origin/gh/wconstab/447/base 2025-09-07T07:36:00.1718486Z * [new branch] gh/wconstab/447/head -> origin/gh/wconstab/447/head 2025-09-07T07:36:00.1718882Z * [new branch] gh/wconstab/447/orig -> origin/gh/wconstab/447/orig 2025-09-07T07:36:00.1720070Z * [new branch] gh/weifengpy/27/base -> origin/gh/weifengpy/27/base 2025-09-07T07:36:00.1720471Z * [new branch] gh/weifengpy/27/head -> origin/gh/weifengpy/27/head 2025-09-07T07:36:00.1721139Z * [new branch] gh/weifengpy/27/orig -> origin/gh/weifengpy/27/orig 2025-09-07T07:36:00.1722024Z * [new branch] gh/weifengpy/30/base -> origin/gh/weifengpy/30/base 2025-09-07T07:36:00.1722415Z * [new branch] gh/weifengpy/30/head -> origin/gh/weifengpy/30/head 2025-09-07T07:36:00.1723010Z * [new branch] gh/weifengpy/30/orig -> origin/gh/weifengpy/30/orig 2025-09-07T07:36:00.1724126Z * [new branch] gh/williamwen42/196/base -> origin/gh/williamwen42/196/base 2025-09-07T07:36:00.1724546Z * [new branch] gh/williamwen42/196/head -> origin/gh/williamwen42/196/head 2025-09-07T07:36:00.1725233Z * [new branch] gh/williamwen42/196/orig -> origin/gh/williamwen42/196/orig 2025-09-07T07:36:00.1726038Z * [new branch] gh/williamwen42/250/base -> origin/gh/williamwen42/250/base 2025-09-07T07:36:00.1726481Z * [new branch] gh/williamwen42/250/head -> origin/gh/williamwen42/250/head 2025-09-07T07:36:00.1727116Z * [new branch] gh/williamwen42/250/orig -> origin/gh/williamwen42/250/orig 2025-09-07T07:36:00.1727920Z * [new branch] gh/williamwen42/258/base -> origin/gh/williamwen42/258/base 2025-09-07T07:36:00.1728591Z * [new branch] gh/williamwen42/258/head -> origin/gh/williamwen42/258/head 2025-09-07T07:36:00.1729165Z * [new branch] gh/williamwen42/258/orig -> origin/gh/williamwen42/258/orig 2025-09-07T07:36:00.1729867Z * [new branch] gh/williamwen42/266/base -> origin/gh/williamwen42/266/base 2025-09-07T07:36:00.1730433Z * [new branch] gh/williamwen42/266/head -> origin/gh/williamwen42/266/head 2025-09-07T07:36:00.1731084Z * [new branch] gh/williamwen42/266/orig -> origin/gh/williamwen42/266/orig 2025-09-07T07:36:00.1731903Z * [new branch] gh/williamwen42/267/base -> origin/gh/williamwen42/267/base 2025-09-07T07:36:00.1732453Z * [new branch] gh/williamwen42/267/head -> origin/gh/williamwen42/267/head 2025-09-07T07:36:00.1733008Z * [new branch] gh/williamwen42/267/orig -> origin/gh/williamwen42/267/orig 2025-09-07T07:36:00.1733811Z * [new branch] gh/williamwen42/270/base -> origin/gh/williamwen42/270/base 2025-09-07T07:36:00.1734371Z * [new branch] gh/williamwen42/270/head -> origin/gh/williamwen42/270/head 2025-09-07T07:36:00.1734971Z * [new branch] gh/williamwen42/270/orig -> origin/gh/williamwen42/270/orig 2025-09-07T07:36:00.1735750Z * [new branch] gh/williamwen42/271/base -> origin/gh/williamwen42/271/base 2025-09-07T07:36:00.1736361Z * [new branch] gh/williamwen42/271/head -> origin/gh/williamwen42/271/head 2025-09-07T07:36:00.1737024Z * [new branch] gh/williamwen42/271/orig -> origin/gh/williamwen42/271/orig 2025-09-07T07:36:00.1737691Z * [new branch] gh/williamwen42/272/base -> origin/gh/williamwen42/272/base 2025-09-07T07:36:00.1738239Z * [new branch] gh/williamwen42/272/head -> origin/gh/williamwen42/272/head 2025-09-07T07:36:00.1738824Z * [new branch] gh/williamwen42/272/orig -> origin/gh/williamwen42/272/orig 2025-09-07T07:36:00.1739506Z * [new branch] gh/williamwen42/274/base -> origin/gh/williamwen42/274/base 2025-09-07T07:36:00.1739961Z * [new branch] gh/williamwen42/274/head -> origin/gh/williamwen42/274/head 2025-09-07T07:36:00.1740670Z * [new branch] gh/williamwen42/274/orig -> origin/gh/williamwen42/274/orig 2025-09-07T07:36:00.1741473Z * [new branch] gh/williamwen42/275/base -> origin/gh/williamwen42/275/base 2025-09-07T07:36:00.1741860Z * [new branch] gh/williamwen42/275/head -> origin/gh/williamwen42/275/head 2025-09-07T07:36:00.1742631Z * [new branch] gh/williamwen42/276/base -> origin/gh/williamwen42/276/base 2025-09-07T07:36:00.1743057Z * [new branch] gh/williamwen42/276/head -> origin/gh/williamwen42/276/head 2025-09-07T07:36:00.1743676Z * [new branch] gh/williamwen42/276/orig -> origin/gh/williamwen42/276/orig 2025-09-07T07:36:00.1744601Z * [new branch] gh/williamwen42/277/base -> origin/gh/williamwen42/277/base 2025-09-07T07:36:00.1745048Z * [new branch] gh/williamwen42/277/head -> origin/gh/williamwen42/277/head 2025-09-07T07:36:00.1745639Z * [new branch] gh/williamwen42/277/orig -> origin/gh/williamwen42/277/orig 2025-09-07T07:36:00.1746398Z * [new branch] gh/williamwen42/278/base -> origin/gh/williamwen42/278/base 2025-09-07T07:36:00.1746961Z * [new branch] gh/williamwen42/278/head -> origin/gh/williamwen42/278/head 2025-09-07T07:36:00.1747515Z * [new branch] gh/williamwen42/278/orig -> origin/gh/williamwen42/278/orig 2025-09-07T07:36:00.1748292Z * [new branch] gh/williamwen42/279/base -> origin/gh/williamwen42/279/base 2025-09-07T07:36:00.1748924Z * [new branch] gh/williamwen42/279/head -> origin/gh/williamwen42/279/head 2025-09-07T07:36:00.1749511Z * [new branch] gh/williamwen42/279/orig -> origin/gh/williamwen42/279/orig 2025-09-07T07:36:00.1750247Z * [new branch] gh/williamwen42/280/base -> origin/gh/williamwen42/280/base 2025-09-07T07:36:00.1750839Z * [new branch] gh/williamwen42/280/head -> origin/gh/williamwen42/280/head 2025-09-07T07:36:00.1751204Z * [new branch] gh/williamwen42/280/orig -> origin/gh/williamwen42/280/orig 2025-09-07T07:36:00.1752155Z * [new branch] gh/williamwen42/281/base -> origin/gh/williamwen42/281/base 2025-09-07T07:36:00.1752719Z * [new branch] gh/williamwen42/281/head -> origin/gh/williamwen42/281/head 2025-09-07T07:36:00.1753120Z * [new branch] gh/williamwen42/281/orig -> origin/gh/williamwen42/281/orig 2025-09-07T07:36:00.1754099Z * [new branch] gh/williamwen42/282/base -> origin/gh/williamwen42/282/base 2025-09-07T07:36:00.1754554Z * [new branch] gh/williamwen42/282/head -> origin/gh/williamwen42/282/head 2025-09-07T07:36:00.1755157Z * [new branch] gh/williamwen42/282/orig -> origin/gh/williamwen42/282/orig 2025-09-07T07:36:00.1756023Z * [new branch] gh/williamwen42/283/base -> origin/gh/williamwen42/283/base 2025-09-07T07:36:00.1756616Z * [new branch] gh/williamwen42/283/head -> origin/gh/williamwen42/283/head 2025-09-07T07:36:00.1757061Z * [new branch] gh/williamwen42/283/orig -> origin/gh/williamwen42/283/orig 2025-09-07T07:36:00.1758103Z * [new branch] gh/williamwen42/284/base -> origin/gh/williamwen42/284/base 2025-09-07T07:36:00.1758559Z * [new branch] gh/williamwen42/284/head -> origin/gh/williamwen42/284/head 2025-09-07T07:36:00.1759283Z * [new branch] gh/williamwen42/284/orig -> origin/gh/williamwen42/284/orig 2025-09-07T07:36:00.1760035Z * [new branch] gh/williamwen42/285/base -> origin/gh/williamwen42/285/base 2025-09-07T07:36:00.1760474Z * [new branch] gh/williamwen42/285/head -> origin/gh/williamwen42/285/head 2025-09-07T07:36:00.1761137Z * [new branch] gh/williamwen42/285/orig -> origin/gh/williamwen42/285/orig 2025-09-07T07:36:00.1761755Z * [new branch] gh/williamwen42/286/base -> origin/gh/williamwen42/286/base 2025-09-07T07:36:00.1762178Z * [new branch] gh/williamwen42/286/head -> origin/gh/williamwen42/286/head 2025-09-07T07:36:00.1762797Z * [new branch] gh/williamwen42/286/orig -> origin/gh/williamwen42/286/orig 2025-09-07T07:36:00.1764733Z * [new branch] gh/williamwen42/287/base -> origin/gh/williamwen42/287/base 2025-09-07T07:36:00.1765444Z * [new branch] gh/williamwen42/287/head -> origin/gh/williamwen42/287/head 2025-09-07T07:36:00.1765578Z * [new branch] gh/williamwen42/287/orig -> origin/gh/williamwen42/287/orig 2025-09-07T07:36:00.1765722Z * [new branch] gh/williamwen42/288/base -> origin/gh/williamwen42/288/base 2025-09-07T07:36:00.1766282Z * [new branch] gh/williamwen42/288/head -> origin/gh/williamwen42/288/head 2025-09-07T07:36:00.1767127Z * [new branch] gh/williamwen42/288/orig -> origin/gh/williamwen42/288/orig 2025-09-07T07:36:00.1768015Z * [new branch] gh/williamwen42/289/base -> origin/gh/williamwen42/289/base 2025-09-07T07:36:00.1768460Z * [new branch] gh/williamwen42/289/head -> origin/gh/williamwen42/289/head 2025-09-07T07:36:00.1769300Z * [new branch] gh/williamwen42/289/orig -> origin/gh/williamwen42/289/orig 2025-09-07T07:36:00.1770476Z * [new branch] gh/wychi/1/base -> origin/gh/wychi/1/base 2025-09-07T07:36:00.1771148Z * [new branch] gh/wychi/1/head -> origin/gh/wychi/1/head 2025-09-07T07:36:00.1771758Z * [new branch] gh/wychi/1/orig -> origin/gh/wychi/1/orig 2025-09-07T07:36:00.1772735Z * [new branch] gh/xmfan/169/base -> origin/gh/xmfan/169/base 2025-09-07T07:36:00.1773280Z * [new branch] gh/xmfan/169/head -> origin/gh/xmfan/169/head 2025-09-07T07:36:00.1773954Z * [new branch] gh/xmfan/170/base -> origin/gh/xmfan/170/base 2025-09-07T07:36:00.1774350Z * [new branch] gh/xmfan/170/head -> origin/gh/xmfan/170/head 2025-09-07T07:36:00.1775321Z * [new branch] gh/xmfan/18/base -> origin/gh/xmfan/18/base 2025-09-07T07:36:00.1775711Z * [new branch] gh/xmfan/18/head -> origin/gh/xmfan/18/head 2025-09-07T07:36:00.1776520Z * [new branch] gh/xmfan/229/base -> origin/gh/xmfan/229/base 2025-09-07T07:36:00.1777060Z * [new branch] gh/xmfan/229/head -> origin/gh/xmfan/229/head 2025-09-07T07:36:00.1777603Z * [new branch] gh/xmfan/229/orig -> origin/gh/xmfan/229/orig 2025-09-07T07:36:00.1778317Z * [new branch] gh/xmfan/237/base -> origin/gh/xmfan/237/base 2025-09-07T07:36:00.1778752Z * [new branch] gh/xmfan/237/head -> origin/gh/xmfan/237/head 2025-09-07T07:36:00.1779650Z * [new branch] gh/xmfan/237/orig -> origin/gh/xmfan/237/orig 2025-09-07T07:36:00.1780245Z * [new branch] gh/xmfan/244/base -> origin/gh/xmfan/244/base 2025-09-07T07:36:00.1780673Z * [new branch] gh/xmfan/244/head -> origin/gh/xmfan/244/head 2025-09-07T07:36:00.1781285Z * [new branch] gh/xmfan/244/orig -> origin/gh/xmfan/244/orig 2025-09-07T07:36:00.1782059Z * [new branch] gh/xmfan/246/base -> origin/gh/xmfan/246/base 2025-09-07T07:36:00.1782488Z * [new branch] gh/xmfan/246/head -> origin/gh/xmfan/246/head 2025-09-07T07:36:00.1783079Z * [new branch] gh/xmfan/246/orig -> origin/gh/xmfan/246/orig 2025-09-07T07:36:00.1783870Z * [new branch] gh/xmfan/253/base -> origin/gh/xmfan/253/base 2025-09-07T07:36:00.1784252Z * [new branch] gh/xmfan/253/head -> origin/gh/xmfan/253/head 2025-09-07T07:36:00.1784905Z * [new branch] gh/xmfan/253/orig -> origin/gh/xmfan/253/orig 2025-09-07T07:36:00.1785614Z * [new branch] gh/xmfan/254/base -> origin/gh/xmfan/254/base 2025-09-07T07:36:00.1786175Z * [new branch] gh/xmfan/254/head -> origin/gh/xmfan/254/head 2025-09-07T07:36:00.1786580Z * [new branch] gh/xmfan/254/orig -> origin/gh/xmfan/254/orig 2025-09-07T07:36:00.1787364Z * [new branch] gh/xmfan/260/base -> origin/gh/xmfan/260/base 2025-09-07T07:36:00.1787786Z * [new branch] gh/xmfan/260/head -> origin/gh/xmfan/260/head 2025-09-07T07:36:00.1788527Z * [new branch] gh/xmfan/260/orig -> origin/gh/xmfan/260/orig 2025-09-07T07:36:00.1789603Z * [new branch] gh/xmfan/262/base -> origin/gh/xmfan/262/base 2025-09-07T07:36:00.1790229Z * [new branch] gh/xmfan/262/head -> origin/gh/xmfan/262/head 2025-09-07T07:36:00.1790618Z * [new branch] gh/xmfan/262/orig -> origin/gh/xmfan/262/orig 2025-09-07T07:36:00.1791527Z * [new branch] gh/xmfan/263/base -> origin/gh/xmfan/263/base 2025-09-07T07:36:00.1791891Z * [new branch] gh/xmfan/263/head -> origin/gh/xmfan/263/head 2025-09-07T07:36:00.1792536Z * [new branch] gh/xmfan/263/orig -> origin/gh/xmfan/263/orig 2025-09-07T07:36:00.1793262Z * [new branch] gh/xmfan/264/base -> origin/gh/xmfan/264/base 2025-09-07T07:36:00.1793869Z * [new branch] gh/xmfan/264/head -> origin/gh/xmfan/264/head 2025-09-07T07:36:00.1794194Z * [new branch] gh/xmfan/264/orig -> origin/gh/xmfan/264/orig 2025-09-07T07:36:00.1795013Z * [new branch] gh/xmfan/274/base -> origin/gh/xmfan/274/base 2025-09-07T07:36:00.1795457Z * [new branch] gh/xmfan/274/head -> origin/gh/xmfan/274/head 2025-09-07T07:36:00.1796066Z * [new branch] gh/xmfan/274/orig -> origin/gh/xmfan/274/orig 2025-09-07T07:36:00.1796874Z * [new branch] gh/xmfan/276/base -> origin/gh/xmfan/276/base 2025-09-07T07:36:00.1797496Z * [new branch] gh/xmfan/276/head -> origin/gh/xmfan/276/head 2025-09-07T07:36:00.1798035Z * [new branch] gh/xmfan/276/orig -> origin/gh/xmfan/276/orig 2025-09-07T07:36:00.1798722Z * [new branch] gh/xmfan/277/base -> origin/gh/xmfan/277/base 2025-09-07T07:36:00.1799330Z * [new branch] gh/xmfan/277/head -> origin/gh/xmfan/277/head 2025-09-07T07:36:00.1799674Z * [new branch] gh/xmfan/277/orig -> origin/gh/xmfan/277/orig 2025-09-07T07:36:00.1800465Z * [new branch] gh/xmfan/278/base -> origin/gh/xmfan/278/base 2025-09-07T07:36:00.1800910Z * [new branch] gh/xmfan/278/head -> origin/gh/xmfan/278/head 2025-09-07T07:36:00.1801592Z * [new branch] gh/xmfan/278/orig -> origin/gh/xmfan/278/orig 2025-09-07T07:36:00.1802524Z * [new branch] gh/xmfan/279/base -> origin/gh/xmfan/279/base 2025-09-07T07:36:00.1802920Z * [new branch] gh/xmfan/279/head -> origin/gh/xmfan/279/head 2025-09-07T07:36:00.1803517Z * [new branch] gh/xmfan/279/orig -> origin/gh/xmfan/279/orig 2025-09-07T07:36:00.1804345Z * [new branch] gh/xmfan/280/base -> origin/gh/xmfan/280/base 2025-09-07T07:36:00.1804938Z * [new branch] gh/xmfan/280/head -> origin/gh/xmfan/280/head 2025-09-07T07:36:00.1805306Z * [new branch] gh/xmfan/280/orig -> origin/gh/xmfan/280/orig 2025-09-07T07:36:00.1806083Z * [new branch] gh/xmfan/281/base -> origin/gh/xmfan/281/base 2025-09-07T07:36:00.1806748Z * [new branch] gh/xmfan/281/head -> origin/gh/xmfan/281/head 2025-09-07T07:36:00.1807309Z * [new branch] gh/xmfan/281/orig -> origin/gh/xmfan/281/orig 2025-09-07T07:36:00.1808019Z * [new branch] gh/xmfan/282/base -> origin/gh/xmfan/282/base 2025-09-07T07:36:00.1808565Z * [new branch] gh/xmfan/282/head -> origin/gh/xmfan/282/head 2025-09-07T07:36:00.1809288Z * [new branch] gh/xmfan/283/base -> origin/gh/xmfan/283/base 2025-09-07T07:36:00.1809824Z * [new branch] gh/xmfan/283/head -> origin/gh/xmfan/283/head 2025-09-07T07:36:00.1810366Z * [new branch] gh/xmfan/283/orig -> origin/gh/xmfan/283/orig 2025-09-07T07:36:00.1811354Z * [new branch] gh/xuanzhang816/14/base -> origin/gh/xuanzhang816/14/base 2025-09-07T07:36:00.1814408Z * [new branch] gh/xuanzhang816/14/head -> origin/gh/xuanzhang816/14/head 2025-09-07T07:36:00.1814884Z * [new branch] gh/xuanzhang816/14/orig -> origin/gh/xuanzhang816/14/orig 2025-09-07T07:36:00.1815759Z * [new branch] gh/xuanzhang816/19/base -> origin/gh/xuanzhang816/19/base 2025-09-07T07:36:00.1816190Z * [new branch] gh/xuanzhang816/19/head -> origin/gh/xuanzhang816/19/head 2025-09-07T07:36:00.1816836Z * [new branch] gh/xuanzhang816/19/orig -> origin/gh/xuanzhang816/19/orig 2025-09-07T07:36:00.1817593Z * [new branch] gh/xuanzhang816/22/base -> origin/gh/xuanzhang816/22/base 2025-09-07T07:36:00.1818078Z * [new branch] gh/xuanzhang816/22/head -> origin/gh/xuanzhang816/22/head 2025-09-07T07:36:00.1818804Z * [new branch] gh/xuanzhang816/22/orig -> origin/gh/xuanzhang816/22/orig 2025-09-07T07:36:00.1819557Z * [new branch] gh/xuanzhang816/23/base -> origin/gh/xuanzhang816/23/base 2025-09-07T07:36:00.1820012Z * [new branch] gh/xuanzhang816/23/head -> origin/gh/xuanzhang816/23/head 2025-09-07T07:36:00.1820598Z * [new branch] gh/xuanzhang816/23/orig -> origin/gh/xuanzhang816/23/orig 2025-09-07T07:36:00.1821401Z * [new branch] gh/xuanzhang816/24/base -> origin/gh/xuanzhang816/24/base 2025-09-07T07:36:00.1821832Z * [new branch] gh/xuanzhang816/24/head -> origin/gh/xuanzhang816/24/head 2025-09-07T07:36:00.1822389Z * [new branch] gh/xuanzhang816/24/orig -> origin/gh/xuanzhang816/24/orig 2025-09-07T07:36:00.1823120Z * [new branch] gh/xuanzhang816/25/base -> origin/gh/xuanzhang816/25/base 2025-09-07T07:36:00.1823531Z * [new branch] gh/xuanzhang816/25/head -> origin/gh/xuanzhang816/25/head 2025-09-07T07:36:00.1824139Z * [new branch] gh/xuanzhang816/25/orig -> origin/gh/xuanzhang816/25/orig 2025-09-07T07:36:00.1825171Z * [new branch] gh/xuanzhang816/26/base -> origin/gh/xuanzhang816/26/base 2025-09-07T07:36:00.1825748Z * [new branch] gh/xuanzhang816/26/head -> origin/gh/xuanzhang816/26/head 2025-09-07T07:36:00.1826174Z * [new branch] gh/xuanzhang816/26/orig -> origin/gh/xuanzhang816/26/orig 2025-09-07T07:36:00.1827243Z * [new branch] gh/yanbing-j/11/base -> origin/gh/yanbing-j/11/base 2025-09-07T07:36:00.1827689Z * [new branch] gh/yanbing-j/11/head -> origin/gh/yanbing-j/11/head 2025-09-07T07:36:00.1828402Z * [new branch] gh/yanbing-j/11/orig -> origin/gh/yanbing-j/11/orig 2025-09-07T07:36:00.1829161Z * [new branch] gh/yanbing-j/12/base -> origin/gh/yanbing-j/12/base 2025-09-07T07:36:00.1829801Z * [new branch] gh/yanbing-j/12/head -> origin/gh/yanbing-j/12/head 2025-09-07T07:36:00.1830210Z * [new branch] gh/yanbing-j/12/orig -> origin/gh/yanbing-j/12/orig 2025-09-07T07:36:00.1831075Z * [new branch] gh/yanbing-j/13/base -> origin/gh/yanbing-j/13/base 2025-09-07T07:36:00.1831515Z * [new branch] gh/yanbing-j/13/head -> origin/gh/yanbing-j/13/head 2025-09-07T07:36:00.1832111Z * [new branch] gh/yanbing-j/13/orig -> origin/gh/yanbing-j/13/orig 2025-09-07T07:36:00.1832787Z * [new branch] gh/yanbing-j/14/base -> origin/gh/yanbing-j/14/base 2025-09-07T07:36:00.1833390Z * [new branch] gh/yanbing-j/14/head -> origin/gh/yanbing-j/14/head 2025-09-07T07:36:00.1833782Z * [new branch] gh/yanbing-j/14/orig -> origin/gh/yanbing-j/14/orig 2025-09-07T07:36:00.1834535Z * [new branch] gh/yanbing-j/15/base -> origin/gh/yanbing-j/15/base 2025-09-07T07:36:00.1835112Z * [new branch] gh/yanbing-j/15/head -> origin/gh/yanbing-j/15/head 2025-09-07T07:36:00.1835982Z * [new branch] gh/yanbing-j/15/orig -> origin/gh/yanbing-j/15/orig 2025-09-07T07:36:00.1836731Z * [new branch] gh/yanbing-j/18/base -> origin/gh/yanbing-j/18/base 2025-09-07T07:36:00.1837316Z * [new branch] gh/yanbing-j/18/head -> origin/gh/yanbing-j/18/head 2025-09-07T07:36:00.1837934Z * [new branch] gh/yanbing-j/18/orig -> origin/gh/yanbing-j/18/orig 2025-09-07T07:36:00.1838659Z * [new branch] gh/yanbing-j/19/base -> origin/gh/yanbing-j/19/base 2025-09-07T07:36:00.1839201Z * [new branch] gh/yanbing-j/19/head -> origin/gh/yanbing-j/19/head 2025-09-07T07:36:00.1839636Z * [new branch] gh/yanbing-j/19/orig -> origin/gh/yanbing-j/19/orig 2025-09-07T07:36:00.1840442Z * [new branch] gh/yanbing-j/20/base -> origin/gh/yanbing-j/20/base 2025-09-07T07:36:00.1840891Z * [new branch] gh/yanbing-j/20/head -> origin/gh/yanbing-j/20/head 2025-09-07T07:36:00.1841488Z * [new branch] gh/yanbing-j/20/orig -> origin/gh/yanbing-j/20/orig 2025-09-07T07:36:00.1842245Z * [new branch] gh/yanbing-j/21/base -> origin/gh/yanbing-j/21/base 2025-09-07T07:36:00.1842735Z * [new branch] gh/yanbing-j/21/head -> origin/gh/yanbing-j/21/head 2025-09-07T07:36:00.1843593Z * [new branch] gh/yanbing-j/22/base -> origin/gh/yanbing-j/22/base 2025-09-07T07:36:00.1844030Z * [new branch] gh/yanbing-j/22/head -> origin/gh/yanbing-j/22/head 2025-09-07T07:36:00.1844659Z * [new branch] gh/yanbing-j/22/orig -> origin/gh/yanbing-j/22/orig 2025-09-07T07:36:00.1845384Z * [new branch] gh/yanbing-j/23/base -> origin/gh/yanbing-j/23/base 2025-09-07T07:36:00.1845819Z * [new branch] gh/yanbing-j/23/head -> origin/gh/yanbing-j/23/head 2025-09-07T07:36:00.1846512Z * [new branch] gh/yanbing-j/23/orig -> origin/gh/yanbing-j/23/orig 2025-09-07T07:36:00.1847227Z * [new branch] gh/yanbing-j/24/base -> origin/gh/yanbing-j/24/base 2025-09-07T07:36:00.1847774Z * [new branch] gh/yanbing-j/24/head -> origin/gh/yanbing-j/24/head 2025-09-07T07:36:00.1848252Z * [new branch] gh/yanbing-j/24/orig -> origin/gh/yanbing-j/24/orig 2025-09-07T07:36:00.1849035Z * [new branch] gh/yanbing-j/25/base -> origin/gh/yanbing-j/25/base 2025-09-07T07:36:00.1849581Z * [new branch] gh/yanbing-j/25/head -> origin/gh/yanbing-j/25/head 2025-09-07T07:36:00.1850123Z * [new branch] gh/yanbing-j/25/orig -> origin/gh/yanbing-j/25/orig 2025-09-07T07:36:00.1850907Z * [new branch] gh/yanbing-j/26/base -> origin/gh/yanbing-j/26/base 2025-09-07T07:36:00.1851701Z * [new branch] gh/yanbing-j/26/head -> origin/gh/yanbing-j/26/head 2025-09-07T07:36:00.1852242Z * [new branch] gh/yanbing-j/26/orig -> origin/gh/yanbing-j/26/orig 2025-09-07T07:36:00.1852976Z * [new branch] gh/yanbing-j/36/base -> origin/gh/yanbing-j/36/base 2025-09-07T07:36:00.1853523Z * [new branch] gh/yanbing-j/36/head -> origin/gh/yanbing-j/36/head 2025-09-07T07:36:00.1853958Z * [new branch] gh/yanbing-j/36/orig -> origin/gh/yanbing-j/36/orig 2025-09-07T07:36:00.1854755Z * [new branch] gh/yanbing-j/37/base -> origin/gh/yanbing-j/37/base 2025-09-07T07:36:00.1855384Z * [new branch] gh/yanbing-j/37/head -> origin/gh/yanbing-j/37/head 2025-09-07T07:36:00.1855828Z * [new branch] gh/yanbing-j/37/orig -> origin/gh/yanbing-j/37/orig 2025-09-07T07:36:00.1856810Z * [new branch] gh/yangw-dev/12/base -> origin/gh/yangw-dev/12/base 2025-09-07T07:36:00.1857271Z * [new branch] gh/yangw-dev/12/head -> origin/gh/yangw-dev/12/head 2025-09-07T07:36:00.1857818Z * [new branch] gh/yangw-dev/12/orig -> origin/gh/yangw-dev/12/orig 2025-09-07T07:36:00.1858699Z * [new branch] gh/yangw-dev/13/base -> origin/gh/yangw-dev/13/base 2025-09-07T07:36:00.1859166Z * [new branch] gh/yangw-dev/13/head -> origin/gh/yangw-dev/13/head 2025-09-07T07:36:00.1859818Z * [new branch] gh/yangw-dev/13/orig -> origin/gh/yangw-dev/13/orig 2025-09-07T07:36:00.1860457Z * [new branch] gh/yangw-dev/14/base -> origin/gh/yangw-dev/14/base 2025-09-07T07:36:00.1860997Z * [new branch] gh/yangw-dev/14/head -> origin/gh/yangw-dev/14/head 2025-09-07T07:36:00.1861437Z * [new branch] gh/yangw-dev/14/orig -> origin/gh/yangw-dev/14/orig 2025-09-07T07:36:00.1862301Z * [new branch] gh/yangw-dev/15/base -> origin/gh/yangw-dev/15/base 2025-09-07T07:36:00.1862748Z * [new branch] gh/yangw-dev/15/head -> origin/gh/yangw-dev/15/head 2025-09-07T07:36:00.1863342Z * [new branch] gh/yangw-dev/15/orig -> origin/gh/yangw-dev/15/orig 2025-09-07T07:36:00.1864121Z * [new branch] gh/yangw-dev/16/base -> origin/gh/yangw-dev/16/base 2025-09-07T07:36:00.1864494Z * [new branch] gh/yangw-dev/16/head -> origin/gh/yangw-dev/16/head 2025-09-07T07:36:00.1865196Z * [new branch] gh/yangw-dev/16/orig -> origin/gh/yangw-dev/16/orig 2025-09-07T07:36:00.1866028Z * [new branch] gh/yangw-dev/17/base -> origin/gh/yangw-dev/17/base 2025-09-07T07:36:00.1866788Z * [new branch] gh/yangw-dev/17/head -> origin/gh/yangw-dev/17/head 2025-09-07T07:36:00.1867373Z * [new branch] gh/yangw-dev/17/orig -> origin/gh/yangw-dev/17/orig 2025-09-07T07:36:00.1868031Z * [new branch] gh/yangw-dev/18/base -> origin/gh/yangw-dev/18/base 2025-09-07T07:36:00.1868473Z * [new branch] gh/yangw-dev/18/head -> origin/gh/yangw-dev/18/head 2025-09-07T07:36:00.1869057Z * [new branch] gh/yangw-dev/18/orig -> origin/gh/yangw-dev/18/orig 2025-09-07T07:36:00.1869819Z * [new branch] gh/yangw-dev/19/base -> origin/gh/yangw-dev/19/base 2025-09-07T07:36:00.1870308Z * [new branch] gh/yangw-dev/19/head -> origin/gh/yangw-dev/19/head 2025-09-07T07:36:00.1870846Z * [new branch] gh/yangw-dev/19/orig -> origin/gh/yangw-dev/19/orig 2025-09-07T07:36:00.1871547Z * [new branch] gh/yangw-dev/20/base -> origin/gh/yangw-dev/20/base 2025-09-07T07:36:00.1872360Z * [new branch] gh/yangw-dev/20/head -> origin/gh/yangw-dev/20/head 2025-09-07T07:36:00.1872902Z * [new branch] gh/yangw-dev/20/orig -> origin/gh/yangw-dev/20/orig 2025-09-07T07:36:00.1873706Z * [new branch] gh/yangw-dev/21/base -> origin/gh/yangw-dev/21/base 2025-09-07T07:36:00.1874201Z * [new branch] gh/yangw-dev/21/head -> origin/gh/yangw-dev/21/head 2025-09-07T07:36:00.1874864Z * [new branch] gh/yangw-dev/21/orig -> origin/gh/yangw-dev/21/orig 2025-09-07T07:36:00.1875567Z * [new branch] gh/yangw-dev/22/base -> origin/gh/yangw-dev/22/base 2025-09-07T07:36:00.1876112Z * [new branch] gh/yangw-dev/22/head -> origin/gh/yangw-dev/22/head 2025-09-07T07:36:00.1876543Z * [new branch] gh/yangw-dev/22/orig -> origin/gh/yangw-dev/22/orig 2025-09-07T07:36:00.1877287Z * [new branch] gh/yangw-dev/23/base -> origin/gh/yangw-dev/23/base 2025-09-07T07:36:00.1877727Z * [new branch] gh/yangw-dev/23/head -> origin/gh/yangw-dev/23/head 2025-09-07T07:36:00.1878306Z * [new branch] gh/yangw-dev/23/orig -> origin/gh/yangw-dev/23/orig 2025-09-07T07:36:00.1879036Z * [new branch] gh/yangw-dev/24/base -> origin/gh/yangw-dev/24/base 2025-09-07T07:36:00.1879560Z * [new branch] gh/yangw-dev/24/head -> origin/gh/yangw-dev/24/head 2025-09-07T07:36:00.1880102Z * [new branch] gh/yangw-dev/24/orig -> origin/gh/yangw-dev/24/orig 2025-09-07T07:36:00.1880869Z * [new branch] gh/yangw-dev/25/base -> origin/gh/yangw-dev/25/base 2025-09-07T07:36:00.1881430Z * [new branch] gh/yangw-dev/25/head -> origin/gh/yangw-dev/25/head 2025-09-07T07:36:00.1881841Z * [new branch] gh/yangw-dev/25/orig -> origin/gh/yangw-dev/25/orig 2025-09-07T07:36:00.1882635Z * [new branch] gh/yangw-dev/26/base -> origin/gh/yangw-dev/26/base 2025-09-07T07:36:00.1883061Z * [new branch] gh/yangw-dev/26/head -> origin/gh/yangw-dev/26/head 2025-09-07T07:36:00.1883743Z * [new branch] gh/yangw-dev/26/orig -> origin/gh/yangw-dev/26/orig 2025-09-07T07:36:00.1884508Z * [new branch] gh/yangw-dev/27/base -> origin/gh/yangw-dev/27/base 2025-09-07T07:36:00.1885040Z * [new branch] gh/yangw-dev/27/head -> origin/gh/yangw-dev/27/head 2025-09-07T07:36:00.1885603Z * [new branch] gh/yangw-dev/27/orig -> origin/gh/yangw-dev/27/orig 2025-09-07T07:36:00.1886526Z * [new branch] gh/ydwu4/233/base -> origin/gh/ydwu4/233/base 2025-09-07T07:36:00.1887117Z * [new branch] gh/ydwu4/233/head -> origin/gh/ydwu4/233/head 2025-09-07T07:36:00.1887731Z * [new branch] gh/ydwu4/233/orig -> origin/gh/ydwu4/233/orig 2025-09-07T07:36:00.1888626Z * [new branch] gh/ydwu4/246/base -> origin/gh/ydwu4/246/base 2025-09-07T07:36:00.1889157Z * [new branch] gh/ydwu4/246/head -> origin/gh/ydwu4/246/head 2025-09-07T07:36:00.1889596Z * [new branch] gh/ydwu4/246/orig -> origin/gh/ydwu4/246/orig 2025-09-07T07:36:00.1890509Z * [new branch] gh/ydwu4/253/base -> origin/gh/ydwu4/253/base 2025-09-07T07:36:00.1891109Z * [new branch] gh/ydwu4/253/head -> origin/gh/ydwu4/253/head 2025-09-07T07:36:00.1891650Z * [new branch] gh/ydwu4/253/orig -> origin/gh/ydwu4/253/orig 2025-09-07T07:36:00.1892414Z * [new branch] gh/ydwu4/255/base -> origin/gh/ydwu4/255/base 2025-09-07T07:36:00.1893184Z * [new branch] gh/ydwu4/255/head -> origin/gh/ydwu4/255/head 2025-09-07T07:36:00.1893477Z * [new branch] gh/ydwu4/255/orig -> origin/gh/ydwu4/255/orig 2025-09-07T07:36:00.1894415Z * [new branch] gh/ydwu4/259/base -> origin/gh/ydwu4/259/base 2025-09-07T07:36:00.1894946Z * [new branch] gh/ydwu4/259/head -> origin/gh/ydwu4/259/head 2025-09-07T07:36:00.1895543Z * [new branch] gh/ydwu4/259/orig -> origin/gh/ydwu4/259/orig 2025-09-07T07:36:00.1896311Z * [new branch] gh/ydwu4/262/base -> origin/gh/ydwu4/262/base 2025-09-07T07:36:00.1896899Z * [new branch] gh/ydwu4/262/head -> origin/gh/ydwu4/262/head 2025-09-07T07:36:00.1897334Z * [new branch] gh/ydwu4/262/orig -> origin/gh/ydwu4/262/orig 2025-09-07T07:36:00.1898162Z * [new branch] gh/ydwu4/263/base -> origin/gh/ydwu4/263/base 2025-09-07T07:36:00.1898690Z * [new branch] gh/ydwu4/263/head -> origin/gh/ydwu4/263/head 2025-09-07T07:36:00.1899225Z * [new branch] gh/ydwu4/263/orig -> origin/gh/ydwu4/263/orig 2025-09-07T07:36:00.1900081Z * [new branch] gh/ydwu4/269/base -> origin/gh/ydwu4/269/base 2025-09-07T07:36:00.1900441Z * [new branch] gh/ydwu4/269/head -> origin/gh/ydwu4/269/head 2025-09-07T07:36:00.1901068Z * [new branch] gh/ydwu4/269/orig -> origin/gh/ydwu4/269/orig 2025-09-07T07:36:00.1901811Z * [new branch] gh/ydwu4/270/base -> origin/gh/ydwu4/270/base 2025-09-07T07:36:00.1902515Z * [new branch] gh/ydwu4/270/head -> origin/gh/ydwu4/270/head 2025-09-07T07:36:00.1903151Z * [new branch] gh/ydwu4/270/orig -> origin/gh/ydwu4/270/orig 2025-09-07T07:36:00.1903889Z * [new branch] gh/ydwu4/272/base -> origin/gh/ydwu4/272/base 2025-09-07T07:36:00.1904503Z * [new branch] gh/ydwu4/272/head -> origin/gh/ydwu4/272/head 2025-09-07T07:36:00.1905046Z * [new branch] gh/ydwu4/272/orig -> origin/gh/ydwu4/272/orig 2025-09-07T07:36:00.1905665Z * [new branch] gh/ydwu4/275/base -> origin/gh/ydwu4/275/base 2025-09-07T07:36:00.1906224Z * [new branch] gh/ydwu4/275/head -> origin/gh/ydwu4/275/head 2025-09-07T07:36:00.1906674Z * [new branch] gh/ydwu4/275/orig -> origin/gh/ydwu4/275/orig 2025-09-07T07:36:00.1907417Z * [new branch] gh/ydwu4/276/base -> origin/gh/ydwu4/276/base 2025-09-07T07:36:00.1907949Z * [new branch] gh/ydwu4/276/head -> origin/gh/ydwu4/276/head 2025-09-07T07:36:00.1908485Z * [new branch] gh/ydwu4/276/orig -> origin/gh/ydwu4/276/orig 2025-09-07T07:36:00.1909333Z * [new branch] gh/ydwu4/279/base -> origin/gh/ydwu4/279/base 2025-09-07T07:36:00.1909990Z * [new branch] gh/ydwu4/279/head -> origin/gh/ydwu4/279/head 2025-09-07T07:36:00.1910593Z * [new branch] gh/ydwu4/279/orig -> origin/gh/ydwu4/279/orig 2025-09-07T07:36:00.1911558Z * [new branch] gh/ydwu4/283/base -> origin/gh/ydwu4/283/base 2025-09-07T07:36:00.1912151Z * [new branch] gh/ydwu4/283/head -> origin/gh/ydwu4/283/head 2025-09-07T07:36:00.1912690Z * [new branch] gh/ydwu4/283/orig -> origin/gh/ydwu4/283/orig 2025-09-07T07:36:00.1913687Z * [new branch] gh/ydwu4/289/base -> origin/gh/ydwu4/289/base 2025-09-07T07:36:00.1914224Z * [new branch] gh/ydwu4/289/head -> origin/gh/ydwu4/289/head 2025-09-07T07:36:00.1914842Z * [new branch] gh/ydwu4/289/orig -> origin/gh/ydwu4/289/orig 2025-09-07T07:36:00.1915699Z * [new branch] gh/ydwu4/290/base -> origin/gh/ydwu4/290/base 2025-09-07T07:36:00.1916131Z * [new branch] gh/ydwu4/290/head -> origin/gh/ydwu4/290/head 2025-09-07T07:36:00.1917100Z * [new branch] gh/ydwu4/290/orig -> origin/gh/ydwu4/290/orig 2025-09-07T07:36:00.1918113Z * [new branch] gh/ydwu4/291/base -> origin/gh/ydwu4/291/base 2025-09-07T07:36:00.1918649Z * [new branch] gh/ydwu4/291/head -> origin/gh/ydwu4/291/head 2025-09-07T07:36:00.1919254Z * [new branch] gh/ydwu4/291/orig -> origin/gh/ydwu4/291/orig 2025-09-07T07:36:00.1920036Z * [new branch] gh/ydwu4/292/base -> origin/gh/ydwu4/292/base 2025-09-07T07:36:00.1920471Z * [new branch] gh/ydwu4/292/head -> origin/gh/ydwu4/292/head 2025-09-07T07:36:00.1921067Z * [new branch] gh/ydwu4/292/orig -> origin/gh/ydwu4/292/orig 2025-09-07T07:36:00.1921742Z * [new branch] gh/ydwu4/293/base -> origin/gh/ydwu4/293/base 2025-09-07T07:36:00.1922374Z * [new branch] gh/ydwu4/293/head -> origin/gh/ydwu4/293/head 2025-09-07T07:36:00.1922913Z * [new branch] gh/ydwu4/293/orig -> origin/gh/ydwu4/293/orig 2025-09-07T07:36:00.1923836Z * [new branch] gh/ydwu4/294/base -> origin/gh/ydwu4/294/base 2025-09-07T07:36:00.1924112Z * [new branch] gh/ydwu4/294/head -> origin/gh/ydwu4/294/head 2025-09-07T07:36:00.1924790Z * [new branch] gh/ydwu4/294/orig -> origin/gh/ydwu4/294/orig 2025-09-07T07:36:00.1925547Z * [new branch] gh/ydwu4/295/base -> origin/gh/ydwu4/295/base 2025-09-07T07:36:00.1926092Z * [new branch] gh/ydwu4/295/head -> origin/gh/ydwu4/295/head 2025-09-07T07:36:00.1926684Z * [new branch] gh/ydwu4/295/orig -> origin/gh/ydwu4/295/orig 2025-09-07T07:36:00.1927381Z * [new branch] gh/ydwu4/296/base -> origin/gh/ydwu4/296/base 2025-09-07T07:36:00.1927821Z * [new branch] gh/ydwu4/296/head -> origin/gh/ydwu4/296/head 2025-09-07T07:36:00.1928431Z * [new branch] gh/ydwu4/296/orig -> origin/gh/ydwu4/296/orig 2025-09-07T07:36:00.1929829Z * [new branch] gh/ydwu4/300/base -> origin/gh/ydwu4/300/base 2025-09-07T07:36:00.1930726Z * [new branch] gh/ydwu4/300/head -> origin/gh/ydwu4/300/head 2025-09-07T07:36:00.1931329Z * [new branch] gh/ydwu4/300/orig -> origin/gh/ydwu4/300/orig 2025-09-07T07:36:00.1932698Z * [new branch] gh/ydwu4/301/base -> origin/gh/ydwu4/301/base 2025-09-07T07:36:00.1933312Z * [new branch] gh/ydwu4/301/head -> origin/gh/ydwu4/301/head 2025-09-07T07:36:00.1933789Z * [new branch] gh/ydwu4/301/orig -> origin/gh/ydwu4/301/orig 2025-09-07T07:36:00.1934576Z * [new branch] gh/ydwu4/302/base -> origin/gh/ydwu4/302/base 2025-09-07T07:36:00.1934997Z * [new branch] gh/ydwu4/302/head -> origin/gh/ydwu4/302/head 2025-09-07T07:36:00.1935974Z * [new branch] gh/ydwu4/302/orig -> origin/gh/ydwu4/302/orig 2025-09-07T07:36:00.1936641Z * [new branch] gh/ydwu4/303/base -> origin/gh/ydwu4/303/base 2025-09-07T07:36:00.1937082Z * [new branch] gh/ydwu4/303/head -> origin/gh/ydwu4/303/head 2025-09-07T07:36:00.1937697Z * [new branch] gh/ydwu4/303/orig -> origin/gh/ydwu4/303/orig 2025-09-07T07:36:00.1938346Z * [new branch] gh/ydwu4/304/base -> origin/gh/ydwu4/304/base 2025-09-07T07:36:00.1938932Z * [new branch] gh/ydwu4/304/head -> origin/gh/ydwu4/304/head 2025-09-07T07:36:00.1939527Z * [new branch] gh/ydwu4/304/orig -> origin/gh/ydwu4/304/orig 2025-09-07T07:36:00.1940391Z * [new branch] gh/ydwu4/305/base -> origin/gh/ydwu4/305/base 2025-09-07T07:36:00.1940809Z * [new branch] gh/ydwu4/305/head -> origin/gh/ydwu4/305/head 2025-09-07T07:36:00.1941478Z * [new branch] gh/ydwu4/305/orig -> origin/gh/ydwu4/305/orig 2025-09-07T07:36:00.1942407Z * [new branch] gh/ydwu4/306/base -> origin/gh/ydwu4/306/base 2025-09-07T07:36:00.1942991Z * [new branch] gh/ydwu4/306/head -> origin/gh/ydwu4/306/head 2025-09-07T07:36:00.1943593Z * [new branch] gh/ydwu4/306/orig -> origin/gh/ydwu4/306/orig 2025-09-07T07:36:00.1944305Z * [new branch] gh/ydwu4/307/base -> origin/gh/ydwu4/307/base 2025-09-07T07:36:00.1944701Z * [new branch] gh/ydwu4/307/head -> origin/gh/ydwu4/307/head 2025-09-07T07:36:00.1945349Z * [new branch] gh/ydwu4/307/orig -> origin/gh/ydwu4/307/orig 2025-09-07T07:36:00.1946107Z * [new branch] gh/ydwu4/308/base -> origin/gh/ydwu4/308/base 2025-09-07T07:36:00.1946654Z * [new branch] gh/ydwu4/308/head -> origin/gh/ydwu4/308/head 2025-09-07T07:36:00.1947247Z * [new branch] gh/ydwu4/308/orig -> origin/gh/ydwu4/308/orig 2025-09-07T07:36:00.1947973Z * [new branch] gh/ydwu4/309/base -> origin/gh/ydwu4/309/base 2025-09-07T07:36:00.1948387Z * [new branch] gh/ydwu4/309/head -> origin/gh/ydwu4/309/head 2025-09-07T07:36:00.1949069Z * [new branch] gh/ydwu4/309/orig -> origin/gh/ydwu4/309/orig 2025-09-07T07:36:00.1949844Z * [new branch] gh/ydwu4/310/base -> origin/gh/ydwu4/310/base 2025-09-07T07:36:00.1950471Z * [new branch] gh/ydwu4/310/head -> origin/gh/ydwu4/310/head 2025-09-07T07:36:00.1951063Z * [new branch] gh/ydwu4/310/orig -> origin/gh/ydwu4/310/orig 2025-09-07T07:36:00.1951889Z * [new branch] gh/ydwu4/311/base -> origin/gh/ydwu4/311/base 2025-09-07T07:36:00.1952479Z * [new branch] gh/ydwu4/311/head -> origin/gh/ydwu4/311/head 2025-09-07T07:36:00.1952882Z * [new branch] gh/ydwu4/311/orig -> origin/gh/ydwu4/311/orig 2025-09-07T07:36:00.1953673Z * [new branch] gh/ydwu4/312/base -> origin/gh/ydwu4/312/base 2025-09-07T07:36:00.1954215Z * [new branch] gh/ydwu4/312/head -> origin/gh/ydwu4/312/head 2025-09-07T07:36:00.1954749Z * [new branch] gh/ydwu4/312/orig -> origin/gh/ydwu4/312/orig 2025-09-07T07:36:00.1955711Z * [new branch] gh/ydwu4/313/base -> origin/gh/ydwu4/313/base 2025-09-07T07:36:00.1956245Z * [new branch] gh/ydwu4/313/head -> origin/gh/ydwu4/313/head 2025-09-07T07:36:00.1956740Z * [new branch] gh/ydwu4/313/orig -> origin/gh/ydwu4/313/orig 2025-09-07T07:36:00.1957947Z * [new branch] gh/ydwu4/314/base -> origin/gh/ydwu4/314/base 2025-09-07T07:36:00.1958659Z * [new branch] gh/ydwu4/314/head -> origin/gh/ydwu4/314/head 2025-09-07T07:36:00.1959252Z * [new branch] gh/ydwu4/314/orig -> origin/gh/ydwu4/314/orig 2025-09-07T07:36:00.1960132Z * [new branch] gh/ydwu4/315/base -> origin/gh/ydwu4/315/base 2025-09-07T07:36:00.1960541Z * [new branch] gh/ydwu4/315/head -> origin/gh/ydwu4/315/head 2025-09-07T07:36:00.1961153Z * [new branch] gh/ydwu4/315/orig -> origin/gh/ydwu4/315/orig 2025-09-07T07:36:00.1962126Z * [new branch] gh/ydwu4/316/base -> origin/gh/ydwu4/316/base 2025-09-07T07:36:00.1962738Z * [new branch] gh/ydwu4/316/head -> origin/gh/ydwu4/316/head 2025-09-07T07:36:00.1963180Z * [new branch] gh/ydwu4/316/orig -> origin/gh/ydwu4/316/orig 2025-09-07T07:36:00.1964096Z * [new branch] gh/ydwu4/317/base -> origin/gh/ydwu4/317/base 2025-09-07T07:36:00.1964412Z * [new branch] gh/ydwu4/317/head -> origin/gh/ydwu4/317/head 2025-09-07T07:36:00.1965101Z * [new branch] gh/ydwu4/317/orig -> origin/gh/ydwu4/317/orig 2025-09-07T07:36:00.1965776Z * [new branch] gh/ydwu4/318/base -> origin/gh/ydwu4/318/base 2025-09-07T07:36:00.1966360Z * [new branch] gh/ydwu4/318/head -> origin/gh/ydwu4/318/head 2025-09-07T07:36:00.1966964Z * [new branch] gh/ydwu4/318/orig -> origin/gh/ydwu4/318/orig 2025-09-07T07:36:00.1967650Z * [new branch] gh/ydwu4/319/base -> origin/gh/ydwu4/319/base 2025-09-07T07:36:00.1968066Z * [new branch] gh/ydwu4/319/head -> origin/gh/ydwu4/319/head 2025-09-07T07:36:00.1972604Z * [new branch] gh/ydwu4/319/orig -> origin/gh/ydwu4/319/orig 2025-09-07T07:36:00.1973463Z * [new branch] gh/ydwu4/320/base -> origin/gh/ydwu4/320/base 2025-09-07T07:36:00.1974030Z * [new branch] gh/ydwu4/320/head -> origin/gh/ydwu4/320/head 2025-09-07T07:36:00.1974413Z * [new branch] gh/ydwu4/320/orig -> origin/gh/ydwu4/320/orig 2025-09-07T07:36:00.1975291Z * [new branch] gh/ydwu4/321/base -> origin/gh/ydwu4/321/base 2025-09-07T07:36:00.1975825Z * [new branch] gh/ydwu4/321/head -> origin/gh/ydwu4/321/head 2025-09-07T07:36:00.1976368Z * [new branch] gh/ydwu4/321/orig -> origin/gh/ydwu4/321/orig 2025-09-07T07:36:00.1977106Z * [new branch] gh/ydwu4/322/base -> origin/gh/ydwu4/322/base 2025-09-07T07:36:00.1977591Z * [new branch] gh/ydwu4/322/head -> origin/gh/ydwu4/322/head 2025-09-07T07:36:00.1978246Z * [new branch] gh/ydwu4/322/orig -> origin/gh/ydwu4/322/orig 2025-09-07T07:36:00.1978988Z * [new branch] gh/ydwu4/323/base -> origin/gh/ydwu4/323/base 2025-09-07T07:36:00.1979588Z * [new branch] gh/ydwu4/323/head -> origin/gh/ydwu4/323/head 2025-09-07T07:36:00.1980058Z * [new branch] gh/ydwu4/323/orig -> origin/gh/ydwu4/323/orig 2025-09-07T07:36:00.1980875Z * [new branch] gh/ydwu4/324/base -> origin/gh/ydwu4/324/base 2025-09-07T07:36:00.1981422Z * [new branch] gh/ydwu4/324/head -> origin/gh/ydwu4/324/head 2025-09-07T07:36:00.1981854Z * [new branch] gh/ydwu4/324/orig -> origin/gh/ydwu4/324/orig 2025-09-07T07:36:00.1982878Z * [new branch] gh/yf225/133/base -> origin/gh/yf225/133/base 2025-09-07T07:36:00.1983417Z * [new branch] gh/yf225/133/head -> origin/gh/yf225/133/head 2025-09-07T07:36:00.1984643Z * [new branch] gh/yf225/171/base -> origin/gh/yf225/171/base 2025-09-07T07:36:00.1985342Z * [new branch] gh/yf225/171/head -> origin/gh/yf225/171/head 2025-09-07T07:36:00.1985944Z * [new branch] gh/yf225/171/orig -> origin/gh/yf225/171/orig 2025-09-07T07:36:00.1986741Z * [new branch] gh/yf225/172/base -> origin/gh/yf225/172/base 2025-09-07T07:36:00.1987133Z * [new branch] gh/yf225/172/head -> origin/gh/yf225/172/head 2025-09-07T07:36:00.1987700Z * [new branch] gh/yf225/172/orig -> origin/gh/yf225/172/orig 2025-09-07T07:36:00.1988461Z * [new branch] gh/yf225/93/base -> origin/gh/yf225/93/base 2025-09-07T07:36:00.1988993Z * [new branch] gh/yf225/93/head -> origin/gh/yf225/93/head 2025-09-07T07:36:00.1990391Z * [new branch] gh/yifuwang/152/base -> origin/gh/yifuwang/152/base 2025-09-07T07:36:00.1991102Z * [new branch] gh/yifuwang/152/head -> origin/gh/yifuwang/152/head 2025-09-07T07:36:00.1991750Z * [new branch] gh/yifuwang/152/orig -> origin/gh/yifuwang/152/orig 2025-09-07T07:36:00.1992739Z * [new branch] gh/yifuwang/195/base -> origin/gh/yifuwang/195/base 2025-09-07T07:36:00.1993302Z * [new branch] gh/yifuwang/195/head -> origin/gh/yifuwang/195/head 2025-09-07T07:36:00.1993916Z * [new branch] gh/yifuwang/195/orig -> origin/gh/yifuwang/195/orig 2025-09-07T07:36:00.1995185Z * [new branch] gh/yiming0416/1/base -> origin/gh/yiming0416/1/base 2025-09-07T07:36:00.1995827Z * [new branch] gh/yiming0416/1/head -> origin/gh/yiming0416/1/head 2025-09-07T07:36:00.1996475Z * [new branch] gh/yiming0416/2/base -> origin/gh/yiming0416/2/base 2025-09-07T07:36:00.1996858Z * [new branch] gh/yiming0416/2/head -> origin/gh/yiming0416/2/head 2025-09-07T07:36:00.1997825Z * [new branch] gh/ysiraichi/79/base -> origin/gh/ysiraichi/79/base 2025-09-07T07:36:00.1998381Z * [new branch] gh/ysiraichi/79/head -> origin/gh/ysiraichi/79/head 2025-09-07T07:36:00.1999144Z * [new branch] gh/ysiraichi/79/orig -> origin/gh/ysiraichi/79/orig 2025-09-07T07:36:00.1999850Z * [new branch] gh/ysiraichi/88/base -> origin/gh/ysiraichi/88/base 2025-09-07T07:36:00.2000393Z * [new branch] gh/ysiraichi/88/head -> origin/gh/ysiraichi/88/head 2025-09-07T07:36:00.2001033Z * [new branch] gh/ysiraichi/88/orig -> origin/gh/ysiraichi/88/orig 2025-09-07T07:36:00.2002001Z * [new branch] gh/zhxchen17/25/base -> origin/gh/zhxchen17/25/base 2025-09-07T07:36:00.2002428Z * [new branch] gh/zhxchen17/25/head -> origin/gh/zhxchen17/25/head 2025-09-07T07:36:00.2003032Z * [new branch] gh/zhxchen17/25/orig -> origin/gh/zhxchen17/25/orig 2025-09-07T07:36:00.2003927Z * [new branch] gh/zhxchen17/31/base -> origin/gh/zhxchen17/31/base 2025-09-07T07:36:00.2004529Z * [new branch] gh/zhxchen17/31/head -> origin/gh/zhxchen17/31/head 2025-09-07T07:36:00.2005118Z * [new branch] gh/zhxchen17/31/orig -> origin/gh/zhxchen17/31/orig 2025-09-07T07:36:00.2005987Z * [new branch] gh/zhxchen17/34/base -> origin/gh/zhxchen17/34/base 2025-09-07T07:36:00.2006531Z * [new branch] gh/zhxchen17/34/head -> origin/gh/zhxchen17/34/head 2025-09-07T07:36:00.2007221Z * [new branch] gh/zhxchen17/35/base -> origin/gh/zhxchen17/35/base 2025-09-07T07:36:00.2007566Z * [new branch] gh/zhxchen17/35/head -> origin/gh/zhxchen17/35/head 2025-09-07T07:36:00.2008616Z * [new branch] gh/zhxchen17/37/base -> origin/gh/zhxchen17/37/base 2025-09-07T07:36:00.2009215Z * [new branch] gh/zhxchen17/37/head -> origin/gh/zhxchen17/37/head 2025-09-07T07:36:00.2009754Z * [new branch] gh/zhxchen17/37/orig -> origin/gh/zhxchen17/37/orig 2025-09-07T07:36:00.2010630Z * [new branch] gh/zhxchen17/38/base -> origin/gh/zhxchen17/38/base 2025-09-07T07:36:00.2011213Z * [new branch] gh/zhxchen17/38/head -> origin/gh/zhxchen17/38/head 2025-09-07T07:36:00.2011601Z * [new branch] gh/zhxchen17/38/orig -> origin/gh/zhxchen17/38/orig 2025-09-07T07:36:00.2012385Z * [new branch] gh/zhxchen17/39/base -> origin/gh/zhxchen17/39/base 2025-09-07T07:36:00.2012936Z * [new branch] gh/zhxchen17/39/head -> origin/gh/zhxchen17/39/head 2025-09-07T07:36:00.2013484Z * [new branch] gh/zhxchen17/39/orig -> origin/gh/zhxchen17/39/orig 2025-09-07T07:36:00.2014358Z * [new branch] gh/zhxchen17/40/base -> origin/gh/zhxchen17/40/base 2025-09-07T07:36:00.2015211Z * [new branch] gh/zhxchen17/40/head -> origin/gh/zhxchen17/40/head 2025-09-07T07:36:00.2015924Z * [new branch] gh/zhxchen17/40/orig -> origin/gh/zhxchen17/40/orig 2025-09-07T07:36:00.2016747Z * [new branch] gh/zhxchen17/41/base -> origin/gh/zhxchen17/41/base 2025-09-07T07:36:00.2017444Z * [new branch] gh/zhxchen17/41/head -> origin/gh/zhxchen17/41/head 2025-09-07T07:36:00.2018196Z * [new branch] gh/zhxchen17/41/orig -> origin/gh/zhxchen17/41/orig 2025-09-07T07:36:00.2019082Z * [new branch] gh/zhxchen17/42/base -> origin/gh/zhxchen17/42/base 2025-09-07T07:36:00.2019738Z * [new branch] gh/zhxchen17/42/head -> origin/gh/zhxchen17/42/head 2025-09-07T07:36:00.2020480Z * [new branch] gh/zhxchen17/42/orig -> origin/gh/zhxchen17/42/orig 2025-09-07T07:36:00.2021334Z * [new branch] gh/zhxchen17/43/base -> origin/gh/zhxchen17/43/base 2025-09-07T07:36:00.2021884Z * [new branch] gh/zhxchen17/43/head -> origin/gh/zhxchen17/43/head 2025-09-07T07:36:00.2022515Z * [new branch] gh/zhxchen17/43/orig -> origin/gh/zhxchen17/43/orig 2025-09-07T07:36:00.2023309Z * [new branch] gh/zhxchen17/44/base -> origin/gh/zhxchen17/44/base 2025-09-07T07:36:00.2023714Z * [new branch] gh/zhxchen17/44/head -> origin/gh/zhxchen17/44/head 2025-09-07T07:36:00.2024366Z * [new branch] gh/zhxchen17/44/orig -> origin/gh/zhxchen17/44/orig 2025-09-07T07:36:00.2025147Z * [new branch] gh/zhxchen17/45/base -> origin/gh/zhxchen17/45/base 2025-09-07T07:36:00.2025744Z * [new branch] gh/zhxchen17/45/head -> origin/gh/zhxchen17/45/head 2025-09-07T07:36:00.2026423Z * [new branch] gh/zhxchen17/45/orig -> origin/gh/zhxchen17/45/orig 2025-09-07T07:36:00.2027390Z * [new branch] gh/zklaus/10/base -> origin/gh/zklaus/10/base 2025-09-07T07:36:00.2027861Z * [new branch] gh/zklaus/10/head -> origin/gh/zklaus/10/head 2025-09-07T07:36:00.2028495Z * [new branch] gh/zklaus/10/orig -> origin/gh/zklaus/10/orig 2025-09-07T07:36:00.2029175Z * [new branch] gh/zklaus/11/base -> origin/gh/zklaus/11/base 2025-09-07T07:36:00.2029721Z * [new branch] gh/zklaus/11/head -> origin/gh/zklaus/11/head 2025-09-07T07:36:00.2030306Z * [new branch] gh/zklaus/11/orig -> origin/gh/zklaus/11/orig 2025-09-07T07:36:00.2031075Z * [new branch] gh/zklaus/12/base -> origin/gh/zklaus/12/base 2025-09-07T07:36:00.2031483Z * [new branch] gh/zklaus/12/head -> origin/gh/zklaus/12/head 2025-09-07T07:36:00.2032229Z * [new branch] gh/zklaus/12/orig -> origin/gh/zklaus/12/orig 2025-09-07T07:36:00.2032933Z * [new branch] gh/zklaus/14/base -> origin/gh/zklaus/14/base 2025-09-07T07:36:00.2033410Z * [new branch] gh/zklaus/14/head -> origin/gh/zklaus/14/head 2025-09-07T07:36:00.2034002Z * [new branch] gh/zklaus/14/orig -> origin/gh/zklaus/14/orig 2025-09-07T07:36:00.2034850Z * [new branch] gh/zklaus/15/base -> origin/gh/zklaus/15/base 2025-09-07T07:36:00.2035260Z * [new branch] gh/zklaus/15/head -> origin/gh/zklaus/15/head 2025-09-07T07:36:00.2035961Z * [new branch] gh/zklaus/15/orig -> origin/gh/zklaus/15/orig 2025-09-07T07:36:00.2036735Z * [new branch] gh/zklaus/16/base -> origin/gh/zklaus/16/base 2025-09-07T07:36:00.2037285Z * [new branch] gh/zklaus/16/head -> origin/gh/zklaus/16/head 2025-09-07T07:36:00.2037714Z * [new branch] gh/zklaus/16/orig -> origin/gh/zklaus/16/orig 2025-09-07T07:36:00.2038546Z * [new branch] gh/zklaus/17/base -> origin/gh/zklaus/17/base 2025-09-07T07:36:00.2039094Z * [new branch] gh/zklaus/17/head -> origin/gh/zklaus/17/head 2025-09-07T07:36:00.2039729Z * [new branch] gh/zklaus/17/orig -> origin/gh/zklaus/17/orig 2025-09-07T07:36:00.2040389Z * [new branch] gh/zklaus/18/base -> origin/gh/zklaus/18/base 2025-09-07T07:36:00.2040939Z * [new branch] gh/zklaus/18/head -> origin/gh/zklaus/18/head 2025-09-07T07:36:00.2041436Z * [new branch] gh/zklaus/18/orig -> origin/gh/zklaus/18/orig 2025-09-07T07:36:00.2042214Z * [new branch] gh/zklaus/19/base -> origin/gh/zklaus/19/base 2025-09-07T07:36:00.2042755Z * [new branch] gh/zklaus/19/head -> origin/gh/zklaus/19/head 2025-09-07T07:36:00.2043656Z * [new branch] gh/zklaus/19/orig -> origin/gh/zklaus/19/orig 2025-09-07T07:36:00.2044401Z * [new branch] gh/zklaus/20/base -> origin/gh/zklaus/20/base 2025-09-07T07:36:00.2045004Z * [new branch] gh/zklaus/20/head -> origin/gh/zklaus/20/head 2025-09-07T07:36:00.2045598Z * [new branch] gh/zklaus/20/orig -> origin/gh/zklaus/20/orig 2025-09-07T07:36:00.2046346Z * [new branch] gh/zklaus/7/base -> origin/gh/zklaus/7/base 2025-09-07T07:36:00.2046869Z * [new branch] gh/zklaus/7/head -> origin/gh/zklaus/7/head 2025-09-07T07:36:00.2047511Z * [new branch] gh/zklaus/7/orig -> origin/gh/zklaus/7/orig 2025-09-07T07:36:00.2048139Z * [new branch] gh/zklaus/9/base -> origin/gh/zklaus/9/base 2025-09-07T07:36:00.2048675Z * [new branch] gh/zklaus/9/head -> origin/gh/zklaus/9/head 2025-09-07T07:36:00.2049214Z * [new branch] gh/zklaus/9/orig -> origin/gh/zklaus/9/orig 2025-09-07T07:36:00.2050196Z * [new branch] gh/zou3519/1175/base -> origin/gh/zou3519/1175/base 2025-09-07T07:36:00.2050680Z * [new branch] gh/zou3519/1175/head -> origin/gh/zou3519/1175/head 2025-09-07T07:36:00.2051326Z * [new branch] gh/zou3519/1175/orig -> origin/gh/zou3519/1175/orig 2025-09-07T07:36:00.2052089Z * [new branch] gh/zou3519/1177/base -> origin/gh/zou3519/1177/base 2025-09-07T07:36:00.2052634Z * [new branch] gh/zou3519/1177/head -> origin/gh/zou3519/1177/head 2025-09-07T07:36:00.2053267Z * [new branch] gh/zou3519/1177/orig -> origin/gh/zou3519/1177/orig 2025-09-07T07:36:00.2054027Z * [new branch] gh/zou3519/1191/base -> origin/gh/zou3519/1191/base 2025-09-07T07:36:00.2054869Z * [new branch] gh/zou3519/1191/head -> origin/gh/zou3519/1191/head 2025-09-07T07:36:00.2055412Z * [new branch] gh/zou3519/1191/orig -> origin/gh/zou3519/1191/orig 2025-09-07T07:36:00.2056288Z * [new branch] gh/zou3519/1192/base -> origin/gh/zou3519/1192/base 2025-09-07T07:36:00.2056846Z * [new branch] gh/zou3519/1192/head -> origin/gh/zou3519/1192/head 2025-09-07T07:36:00.2057335Z * [new branch] gh/zou3519/1192/orig -> origin/gh/zou3519/1192/orig 2025-09-07T07:36:00.2058105Z * [new branch] gh/zou3519/1193/base -> origin/gh/zou3519/1193/base 2025-09-07T07:36:00.2058664Z * [new branch] gh/zou3519/1193/head -> origin/gh/zou3519/1193/head 2025-09-07T07:36:00.2059063Z * [new branch] gh/zou3519/1193/orig -> origin/gh/zou3519/1193/orig 2025-09-07T07:36:00.2059789Z * [new branch] gh/zou3519/1194/base -> origin/gh/zou3519/1194/base 2025-09-07T07:36:00.2060404Z * [new branch] gh/zou3519/1194/head -> origin/gh/zou3519/1194/head 2025-09-07T07:36:00.2060944Z * [new branch] gh/zou3519/1194/orig -> origin/gh/zou3519/1194/orig 2025-09-07T07:36:00.2061764Z * [new branch] gh/zou3519/1195/base -> origin/gh/zou3519/1195/base 2025-09-07T07:36:00.2062488Z * [new branch] gh/zou3519/1195/head -> origin/gh/zou3519/1195/head 2025-09-07T07:36:00.2062898Z * [new branch] gh/zou3519/1195/orig -> origin/gh/zou3519/1195/orig 2025-09-07T07:36:00.2063686Z * [new branch] gh/zou3519/1196/base -> origin/gh/zou3519/1196/base 2025-09-07T07:36:00.2064294Z * [new branch] gh/zou3519/1196/head -> origin/gh/zou3519/1196/head 2025-09-07T07:36:00.2064896Z * [new branch] gh/zou3519/1196/orig -> origin/gh/zou3519/1196/orig 2025-09-07T07:36:00.2065577Z * [new branch] gh/zou3519/1197/base -> origin/gh/zou3519/1197/base 2025-09-07T07:36:00.2066126Z * [new branch] gh/zou3519/1197/head -> origin/gh/zou3519/1197/head 2025-09-07T07:36:00.2066673Z * [new branch] gh/zou3519/1197/orig -> origin/gh/zou3519/1197/orig 2025-09-07T07:36:00.2067740Z * [new branch] gh/zpcore/1/base -> origin/gh/zpcore/1/base 2025-09-07T07:36:00.2068133Z * [new branch] gh/zpcore/1/head -> origin/gh/zpcore/1/head 2025-09-07T07:36:00.2069250Z * [new branch] gh/zpcore/10/base -> origin/gh/zpcore/10/base 2025-09-07T07:36:00.2069577Z * [new branch] gh/zpcore/10/head -> origin/gh/zpcore/10/head 2025-09-07T07:36:00.2070170Z * [new branch] gh/zpcore/10/orig -> origin/gh/zpcore/10/orig 2025-09-07T07:36:00.2070970Z * [new branch] gh/zpcore/11/base -> origin/gh/zpcore/11/base 2025-09-07T07:36:00.2071598Z * [new branch] gh/zpcore/11/head -> origin/gh/zpcore/11/head 2025-09-07T07:36:00.2072147Z * [new branch] gh/zpcore/11/orig -> origin/gh/zpcore/11/orig 2025-09-07T07:36:00.2073014Z * [new branch] gh/zpcore/12/base -> origin/gh/zpcore/12/base 2025-09-07T07:36:00.2073722Z * [new branch] gh/zpcore/12/head -> origin/gh/zpcore/12/head 2025-09-07T07:36:00.2074385Z * [new branch] gh/zpcore/12/orig -> origin/gh/zpcore/12/orig 2025-09-07T07:36:00.2075578Z * [new branch] gh/zpcore/13/base -> origin/gh/zpcore/13/base 2025-09-07T07:36:00.2076114Z * [new branch] gh/zpcore/13/head -> origin/gh/zpcore/13/head 2025-09-07T07:36:00.2076560Z * [new branch] gh/zpcore/13/orig -> origin/gh/zpcore/13/orig 2025-09-07T07:36:00.2077451Z * [new branch] gh/zpcore/14/base -> origin/gh/zpcore/14/base 2025-09-07T07:36:00.2078369Z * [new branch] gh/zpcore/14/head -> origin/gh/zpcore/14/head 2025-09-07T07:36:00.2079168Z * [new branch] gh/zpcore/2/base -> origin/gh/zpcore/2/base 2025-09-07T07:36:00.2079644Z * [new branch] gh/zpcore/2/head -> origin/gh/zpcore/2/head 2025-09-07T07:36:00.2080378Z * [new branch] gh/zpcore/3/base -> origin/gh/zpcore/3/base 2025-09-07T07:36:00.2080776Z * [new branch] gh/zpcore/3/head -> origin/gh/zpcore/3/head 2025-09-07T07:36:00.2081512Z * [new branch] gh/zpcore/4/base -> origin/gh/zpcore/4/base 2025-09-07T07:36:00.2081901Z * [new branch] gh/zpcore/4/head -> origin/gh/zpcore/4/head 2025-09-07T07:36:00.2082654Z * [new branch] gh/zpcore/5/base -> origin/gh/zpcore/5/base 2025-09-07T07:36:00.2083066Z * [new branch] gh/zpcore/5/head -> origin/gh/zpcore/5/head 2025-09-07T07:36:00.2083937Z * [new branch] gh/zpcore/6/base -> origin/gh/zpcore/6/base 2025-09-07T07:36:00.2084351Z * [new branch] gh/zpcore/6/head -> origin/gh/zpcore/6/head 2025-09-07T07:36:00.2085127Z * [new branch] gh/zpcore/7/base -> origin/gh/zpcore/7/base 2025-09-07T07:36:00.2085562Z * [new branch] gh/zpcore/7/head -> origin/gh/zpcore/7/head 2025-09-07T07:36:00.2086421Z * [new branch] gh/zpcore/8/base -> origin/gh/zpcore/8/base 2025-09-07T07:36:00.2086788Z * [new branch] gh/zpcore/8/head -> origin/gh/zpcore/8/head 2025-09-07T07:36:00.2087640Z * [new branch] google-main -> origin/google-main 2025-09-07T07:36:00.2088485Z * [new branch] guangyey/external_stream -> origin/guangyey/external_stream 2025-09-07T07:36:00.2088902Z * [new branch] guangyey/host_alloc -> origin/guangyey/host_alloc 2025-09-07T07:36:00.2089505Z * [new branch] guangyey/reimport -> origin/guangyey/reimport 2025-09-07T07:36:00.2089901Z * [new branch] guangyey/test_2025 -> origin/guangyey/test_2025 2025-09-07T07:36:00.2090935Z * [new branch] guilhermeleobas/cherry-pick-55d87d9dfd9 -> origin/guilhermeleobas/cherry-pick-55d87d9dfd9 2025-09-07T07:36:00.2091620Z * [new branch] haozhe/bf16-dynamic-shape -> origin/haozhe/bf16-dynamic-shape 2025-09-07T07:36:00.2092226Z * [new branch] hc_baseline -> origin/hc_baseline 2025-09-07T07:36:00.2092830Z * [new branch] hf_update -> origin/hf_update 2025-09-07T07:36:00.2093462Z * [new branch] hhh_decomp_mul -> origin/hhh_decomp_mul 2025-09-07T07:36:00.2094050Z * [new branch] hhh_rand -> origin/hhh_rand 2025-09-07T07:36:00.2094761Z * [new branch] hoy/mmsplitk -> origin/hoy/mmsplitk 2025-09-07T07:36:00.2095192Z * [new branch] hoy/triton-PR3973 -> origin/hoy/triton-PR3973 2025-09-07T07:36:00.2095870Z * [new branch] hoy/triton-coalescing-baseline -> origin/hoy/triton-coalescing-baseline 2025-09-07T07:36:00.2096277Z * [new branch] hoy/triton-coalescing-new -> origin/hoy/triton-coalescing-new 2025-09-07T07:36:00.2096893Z * [new branch] hoy/triton-coalescing-vec -> origin/hoy/triton-coalescing-vec 2025-09-07T07:36:00.2097332Z * [new branch] inductordecompfix -> origin/inductordecompfix 2025-09-07T07:36:00.2097988Z * [new branch] inline -> origin/inline 2025-09-07T07:36:00.2098561Z * [new branch] inlining -> origin/inlining 2025-09-07T07:36:00.2099107Z * [new branch] inlining-ezyang -> origin/inlining-ezyang 2025-09-07T07:36:00.2099666Z * [new branch] install-torchao-0.13.0 -> origin/install-torchao-0.13.0 2025-09-07T07:36:00.2100165Z * [new branch] int8_sdpa -> origin/int8_sdpa 2025-09-07T07:36:00.2100830Z * [new branch] invoke-subgraph -> origin/invoke-subgraph 2025-09-07T07:36:00.2101450Z * [new branch] issue#58739 -> origin/issue#58739 2025-09-07T07:36:00.2102424Z * [new branch] jcaip/test-cusparselt-version-0.6.2 -> origin/jcaip/test-cusparselt-version-0.6.2 2025-09-07T07:36:00.2102862Z * [new branch] jcaip/update-cusparselt-0.6.2 -> origin/jcaip/update-cusparselt-0.6.2 2025-09-07T07:36:00.2103693Z * [new branch] jeanschmidt/disable_rocm_build_tests -> origin/jeanschmidt/disable_rocm_build_tests 2025-09-07T07:36:00.2104265Z * [new branch] jithunnair-amd-patch-1 -> origin/jithunnair-amd-patch-1 2025-09-07T07:36:00.2104986Z * [new branch] jithunnair-amd-patch-2 -> origin/jithunnair-amd-patch-2 2025-09-07T07:36:00.2105667Z * [new branch] justinchu/attention-tests -> origin/justinchu/attention-tests 2025-09-07T07:36:00.2106224Z * [new branch] justinchu/native-qdq -> origin/justinchu/native-qdq 2025-09-07T07:36:00.2107100Z * [new branch] justinchu/ort-122 -> origin/justinchu/ort-122 2025-09-07T07:36:00.2107877Z * [new branch] justinchuby/dynamo-true -> origin/justinchuby/dynamo-true 2025-09-07T07:36:00.2108711Z * [new branch] kainan666/xlf_debug -> origin/kainan666/xlf_debug 2025-09-07T07:36:00.2109294Z * [new branch] kainan_test -> origin/kainan_test 2025-09-07T07:36:00.2109831Z * [new branch] learnablebias -> origin/learnablebias 2025-09-07T07:36:00.2110608Z * [new branch] leslie/test_group_gemm_epilogues -> origin/leslie/test_group_gemm_epilogues 2025-09-07T07:36:00.2111348Z * [new branch] lessw2020/fix_cutlass_cache_error -> origin/lessw2020/fix_cutlass_cache_error 2025-09-07T07:36:00.2112050Z * [new branch] liaoxuan/shm_all_reduce -> origin/liaoxuan/shm_all_reduce 2025-09-07T07:36:00.2112713Z * [new branch] liaoxuan/test_fa_disable_softmax -> origin/liaoxuan/test_fa_disable_softmax 2025-09-07T07:36:00.2113088Z * [new branch] liaoxuan/test_int8_sdpa -> origin/liaoxuan/test_int8_sdpa 2025-09-07T07:36:00.2113695Z * [new branch] lintbuilddocker -> origin/lintbuilddocker 2025-09-07T07:36:00.2114088Z * [new branch] llama4-stable -> origin/llama4-stable 2025-09-07T07:36:00.2114782Z * [new branch] logdetfix -> origin/logdetfix 2025-09-07T07:36:00.2115906Z * [new branch] lts/release/1.8 -> origin/lts/release/1.8 2025-09-07T07:36:00.2116648Z * [new branch] lucaskabela/#94773 -> origin/lucaskabela/#94773 2025-09-07T07:36:00.2117101Z * [new branch] lucaskabela/flop_counter -> origin/lucaskabela/flop_counter 2025-09-07T07:36:00.2117668Z * [new branch] lucaskabela/func_under_decomp -> origin/lucaskabela/func_under_decomp 2025-09-07T07:36:00.2118126Z * [new branch] lucaskabela/functional_in_dynamo -> origin/lucaskabela/functional_in_dynamo 2025-09-07T07:36:00.2118778Z * [new branch] lucaskabela/install_params_as_graph_attr -> origin/lucaskabela/install_params_as_graph_attr 2025-09-07T07:36:00.2119167Z * [new branch] lucaskabela/issue_120648 -> origin/lucaskabela/issue_120648 2025-09-07T07:36:00.2119934Z * [new branch] lucaskabela/misc_typing_dynamo -> origin/lucaskabela/misc_typing_dynamo 2025-09-07T07:36:00.2120734Z * [new branch] lucaskabela/parameters_as_graph_attr -> origin/lucaskabela/parameters_as_graph_attr 2025-09-07T07:36:00.2121244Z * [new branch] lucaskabela/remove_aot_dispatcher_metadata -> origin/lucaskabela/remove_aot_dispatcher_metadata 2025-09-07T07:36:00.2121895Z * [new branch] lucaskabela/rnn_decomp -> origin/lucaskabela/rnn_decomp 2025-09-07T07:36:00.2122342Z * [new branch] lucaskabela/typing_backends -> origin/lucaskabela/typing_backends 2025-09-07T07:36:00.2123012Z * [new branch] lucaskabela/typing_symbolic_convert -> origin/lucaskabela/typing_symbolic_convert 2025-09-07T07:36:00.2123458Z * [new branch] lucaskabela/typing_utils_improvements -> origin/lucaskabela/typing_utils_improvements 2025-09-07T07:36:00.2124192Z * [new branch] main -> origin/main 2025-09-07T07:36:00.2125012Z * [new branch] main-enable-b200-distributed-tests -> origin/main-enable-b200-distributed-tests 2025-09-07T07:36:00.2125551Z * [new branch] malfet-patch-1 -> origin/malfet-patch-1 2025-09-07T07:36:00.2126223Z * [new branch] malfet-patch-12 -> origin/malfet-patch-12 2025-09-07T07:36:00.2126898Z * [new branch] malfet-patch-14 -> origin/malfet-patch-14 2025-09-07T07:36:00.2127888Z * [new branch] malfet-patch-6 -> origin/malfet-patch-6 2025-09-07T07:36:00.2128543Z * [new branch] malfet-patch-8 -> origin/malfet-patch-8 2025-09-07T07:36:00.2129854Z * [new branch] malfet/be-move-more-settings-to-checkout-pytorch -> origin/malfet/be-move-more-settings-to-checkout-pytorch 2025-09-07T07:36:00.2130233Z * [new branch] malfet/delete-upsteam-cuda -> origin/malfet/delete-upsteam-cuda 2025-09-07T07:36:00.2130823Z * [new branch] malfet/mps-implement-col2im -> origin/malfet/mps-implement-col2im 2025-09-07T07:36:00.2131562Z * [new branch] manuel/test-ops-common-allow-mps -> origin/manuel/test-ops-common-allow-mps 2025-09-07T07:36:00.2132220Z * [new branch] metascroy-patch-1 -> origin/metascroy-patch-1 2025-09-07T07:36:00.2133013Z * [new branch] mlazos/S429861-debug -> origin/mlazos/S429861-debug 2025-09-07T07:36:00.2133433Z * [new branch] mlazos/aa -> origin/mlazos/aa 2025-09-07T07:36:00.2134108Z * [new branch] mlazos/arg-renames -> origin/mlazos/arg-renames 2025-09-07T07:36:00.2134699Z * [new branch] mlazos/backup-test-branch -> origin/mlazos/backup-test-branch 2025-09-07T07:36:00.2135063Z * [new branch] mlazos/bad-cudagraphs -> origin/mlazos/bad-cudagraphs 2025-09-07T07:36:00.2135663Z * [new branch] mlazos/baseline -> origin/mlazos/baseline 2025-09-07T07:36:00.2136260Z * [new branch] mlazos/baseline-graph-breaks -> origin/mlazos/baseline-graph-breaks 2025-09-07T07:36:00.2136639Z * [new branch] mlazos/beta-tensor -> origin/mlazos/beta-tensor 2025-09-07T07:36:00.2137389Z * [new branch] mlazos/better-msg -> origin/mlazos/better-msg 2025-09-07T07:36:00.2138103Z * [new branch] mlazos/buffers -> origin/mlazos/buffers 2025-09-07T07:36:00.2138494Z * [new branch] mlazos/buffers2 -> origin/mlazos/buffers2 2025-09-07T07:36:00.2139184Z * [new branch] mlazos/buffers3 -> origin/mlazos/buffers3 2025-09-07T07:36:00.2139978Z * [new branch] mlazos/ck2 -> origin/mlazos/ck2 2025-09-07T07:36:00.2140725Z * [new branch] mlazos/combokernels -> origin/mlazos/combokernels 2025-09-07T07:36:00.2141284Z * [new branch] mlazos/ctx-cleanup -> origin/mlazos/ctx-cleanup 2025-09-07T07:36:00.2141639Z * [new branch] mlazos/cuda-cmd-log -> origin/mlazos/cuda-cmd-log 2025-09-07T07:36:00.2142320Z * [new branch] mlazos/cudagraph-tests -> origin/mlazos/cudagraph-tests 2025-09-07T07:36:00.2142862Z * [new branch] mlazos/cudagraphs-measurement -> origin/mlazos/cudagraphs-measurement 2025-09-07T07:36:00.2143512Z * [new branch] mlazos/cutlass-test -> origin/mlazos/cutlass-test 2025-09-07T07:36:00.2143980Z * [new branch] mlazos/cutlass-topo-bug -> origin/mlazos/cutlass-topo-bug 2025-09-07T07:36:00.2144562Z * [new branch] mlazos/data-gather -> origin/mlazos/data-gather 2025-09-07T07:36:00.2145263Z * [new branch] mlazos/data-ptrs2 -> origin/mlazos/data-ptrs2 2025-09-07T07:36:00.2145705Z * [new branch] mlazos/data-ptrs3 -> origin/mlazos/data-ptrs3 2025-09-07T07:36:00.2146332Z * [new branch] mlazos/dataclass-proxy -> origin/mlazos/dataclass-proxy 2025-09-07T07:36:00.2146739Z * [new branch] mlazos/dc-attrs -> origin/mlazos/dc-attrs 2025-09-07T07:36:00.2147375Z * [new branch] mlazos/dc-helion -> origin/mlazos/dc-helion 2025-09-07T07:36:00.2147746Z * [new branch] mlazos/dict-fix -> origin/mlazos/dict-fix 2025-09-07T07:36:00.2148430Z * [new branch] mlazos/disable-closures -> origin/mlazos/disable-closures 2025-09-07T07:36:00.2149081Z * [new branch] mlazos/disable-tf -> origin/mlazos/disable-tf 2025-09-07T07:36:00.2149445Z * [new branch] mlazos/dupe-fix -> origin/mlazos/dupe-fix 2025-09-07T07:36:00.2150146Z * [new branch] mlazos/dyn-batch -> origin/mlazos/dyn-batch 2025-09-07T07:36:00.2150706Z * [new branch] mlazos/evt -> origin/mlazos/evt 2025-09-07T07:36:00.2151248Z * [new branch] mlazos/exp_disable -> origin/mlazos/exp_disable 2025-09-07T07:36:00.2151916Z * [new branch] mlazos/extract-examples -> origin/mlazos/extract-examples 2025-09-07T07:36:00.2152333Z * [new branch] mlazos/foreach-op -> origin/mlazos/foreach-op 2025-09-07T07:36:00.2152966Z * [new branch] mlazos/fp8 -> origin/mlazos/fp8 2025-09-07T07:36:00.2153381Z * [new branch] mlazos/fp8-bias -> origin/mlazos/fp8-bias 2025-09-07T07:36:00.2154034Z * [new branch] mlazos/fp8-bias-fusion -> origin/mlazos/fp8-bias-fusion 2025-09-07T07:36:00.2154325Z * [new branch] mlazos/fp8-fixes -> origin/mlazos/fp8-fixes 2025-09-07T07:36:00.2154944Z * [new branch] mlazos/freezing -> origin/mlazos/freezing 2025-09-07T07:36:00.2155388Z * [new branch] mlazos/h-comp -> origin/mlazos/h-comp 2025-09-07T07:36:00.2156062Z * [new branch] mlazos/h-comp2 -> origin/mlazos/h-comp2 2025-09-07T07:36:00.2156468Z * [new branch] mlazos/hash-hop -> origin/mlazos/hash-hop 2025-09-07T07:36:00.2157190Z * [new branch] mlazos/hc -> origin/mlazos/hc 2025-09-07T07:36:00.2157788Z * [new branch] mlazos/hc-cycles -> origin/mlazos/hc-cycles 2025-09-07T07:36:00.2158200Z * [new branch] mlazos/hc-fixes -> origin/mlazos/hc-fixes 2025-09-07T07:36:00.2158798Z * [new branch] mlazos/hc-fixes3 -> origin/mlazos/hc-fixes3 2025-09-07T07:36:00.2159225Z * [new branch] mlazos/hc-fixes4 -> origin/mlazos/hc-fixes4 2025-09-07T07:36:00.2159881Z * [new branch] mlazos/hc-hf -> origin/mlazos/hc-hf 2025-09-07T07:36:00.2160305Z * [new branch] mlazos/hc-mut -> origin/mlazos/hc-mut 2025-09-07T07:36:00.2160993Z * [new branch] mlazos/hc10 -> origin/mlazos/hc10 2025-09-07T07:36:00.2161595Z * [new branch] mlazos/hc11 -> origin/mlazos/hc11 2025-09-07T07:36:00.2162000Z * [new branch] mlazos/hc12 -> origin/mlazos/hc12 2025-09-07T07:36:00.2162601Z * [new branch] mlazos/hc13 -> origin/mlazos/hc13 2025-09-07T07:36:00.2163019Z * [new branch] mlazos/hc14 -> origin/mlazos/hc14 2025-09-07T07:36:00.2163623Z * [new branch] mlazos/hc15 -> origin/mlazos/hc15 2025-09-07T07:36:00.2164286Z * [new branch] mlazos/hc2 -> origin/mlazos/hc2 2025-09-07T07:36:00.2164630Z * [new branch] mlazos/hc4 -> origin/mlazos/hc4 2025-09-07T07:36:00.2165313Z * [new branch] mlazos/hc5 -> origin/mlazos/hc5 2025-09-07T07:36:00.2165796Z * [new branch] mlazos/hc6 -> origin/mlazos/hc6 2025-09-07T07:36:00.2166711Z * [new branch] mlazos/hc7 -> origin/mlazos/hc7 2025-09-07T07:36:00.2167023Z * [new branch] mlazos/hc8 -> origin/mlazos/hc8 2025-09-07T07:36:00.2167976Z * [new branch] mlazos/hc9 -> origin/mlazos/hc9 2025-09-07T07:36:00.2168573Z * [new branch] mlazos/hc_baseline2 -> origin/mlazos/hc_baseline2 2025-09-07T07:36:00.2169088Z * [new branch] mlazos/init-per-param -> origin/mlazos/init-per-param 2025-09-07T07:36:00.2169648Z * [new branch] mlazos/init_per_param -> origin/mlazos/init_per_param 2025-09-07T07:36:00.2170272Z * [new branch] mlazos/less-guards -> origin/mlazos/less-guards 2025-09-07T07:36:00.2170714Z * [new branch] mlazos/lr-composibility -> origin/mlazos/lr-composibility 2025-09-07T07:36:00.2171229Z * [new branch] mlazos/main -> origin/mlazos/main 2025-09-07T07:36:00.2171944Z * [new branch] mlazos/main-test-enablement -> origin/mlazos/main-test-enablement 2025-09-07T07:36:00.2172286Z * [new branch] mlazos/main2 -> origin/mlazos/main2 2025-09-07T07:36:00.2172972Z * [new branch] mlazos/mark-static-update -> origin/mlazos/mark-static-update 2025-09-07T07:36:00.2173354Z * [new branch] mlazos/mcg -> origin/mlazos/mcg 2025-09-07T07:36:00.2174049Z * [new branch] mlazos/mcg2 -> origin/mlazos/mcg2 2025-09-07T07:36:00.2174599Z * [new branch] mlazos/meta-guards -> origin/mlazos/meta-guards 2025-09-07T07:36:00.2175385Z * [new branch] mlazos/mlazos/ck2 -> origin/mlazos/mlazos/ck2 2025-09-07T07:36:00.2176011Z * [new branch] mlazos/mlazos/foreach-map-adam -> origin/mlazos/mlazos/foreach-map-adam 2025-09-07T07:36:00.2176445Z * [new branch] mlazos/mlazos/tf-mode-backup -> origin/mlazos/mlazos/tf-mode-backup 2025-09-07T07:36:00.2177058Z * [new branch] mlazos/mod-fix -> origin/mlazos/mod-fix 2025-09-07T07:36:00.2177626Z * [new branch] mlazos/mode-fix -> origin/mlazos/mode-fix 2025-09-07T07:36:00.2178031Z * [new branch] mlazos/more-tests -> origin/mlazos/more-tests 2025-09-07T07:36:00.2178779Z * [new branch] mlazos/no-cpp -> origin/mlazos/no-cpp 2025-09-07T07:36:00.2179423Z * [new branch] mlazos/no-init-group-handling -> origin/mlazos/no-init-group-handling 2025-09-07T07:36:00.2179812Z * [new branch] mlazos/offsets -> origin/mlazos/offsets 2025-09-07T07:36:00.2180439Z * [new branch] mlazos/opt-bench-exp2 -> origin/mlazos/opt-bench-exp2 2025-09-07T07:36:00.2180824Z * [new branch] mlazos/opt-incr -> origin/mlazos/opt-incr 2025-09-07T07:36:00.2181472Z * [new branch] mlazos/proxy-ctors -> origin/mlazos/proxy-ctors 2025-09-07T07:36:00.2181908Z * [new branch] mlazos/quant-fix -> origin/mlazos/quant-fix 2025-09-07T07:36:00.2182670Z * [new branch] mlazos/resnet-fix -> origin/mlazos/resnet-fix 2025-09-07T07:36:00.2183228Z * [new branch] mlazos/revert-inline -> origin/mlazos/revert-inline 2025-09-07T07:36:00.2183790Z * [new branch] mlazos/rm-buf-names -> origin/mlazos/rm-buf-names 2025-09-07T07:36:00.2184120Z * [new branch] mlazos/rm-code -> origin/mlazos/rm-code 2025-09-07T07:36:00.2184741Z * [new branch] mlazos/rm-spam -> origin/mlazos/rm-spam 2025-09-07T07:36:00.2185275Z * [new branch] mlazos/rtp -> origin/mlazos/rtp 2025-09-07T07:36:00.2185838Z * [new branch] mlazos/static-idx-dbg -> origin/mlazos/static-idx-dbg 2025-09-07T07:36:00.2186458Z * [new branch] mlazos/static-inputs-log -> origin/mlazos/static-inputs-log 2025-09-07T07:36:00.2186825Z * [new branch] mlazos/sub-param-fix -> origin/mlazos/sub-param-fix 2025-09-07T07:36:00.2187501Z * [new branch] mlazos/td-fix2 -> origin/mlazos/td-fix2 2025-09-07T07:36:00.2188067Z * [new branch] mlazos/tensor-hasattr2 -> origin/mlazos/tensor-hasattr2 2025-09-07T07:36:00.2188499Z * [new branch] mlazos/test -> origin/mlazos/test 2025-09-07T07:36:00.2189029Z * [new branch] mlazos/tf-mode -> origin/mlazos/tf-mode 2025-09-07T07:36:00.2189702Z * [new branch] mlazos/tf-mode-backup2 -> origin/mlazos/tf-mode-backup2 2025-09-07T07:36:00.2190311Z * [new branch] mlazos/tf-mode-reland -> origin/mlazos/tf-mode-reland 2025-09-07T07:36:00.2190952Z * [new branch] mlazos/tf-mode-reland2 -> origin/mlazos/tf-mode-reland2 2025-09-07T07:36:00.2191560Z * [new branch] mlazos/tf-mode-reland3 -> origin/mlazos/tf-mode-reland3 2025-09-07T07:36:00.2191936Z * [new branch] mlazos/topo-fix -> origin/mlazos/topo-fix 2025-09-07T07:36:00.2192571Z * [new branch] mlazos/triton-no-epi -> origin/mlazos/triton-no-epi 2025-09-07T07:36:00.2192978Z * [new branch] mlazos/tune-proto -> origin/mlazos/tune-proto 2025-09-07T07:36:00.2193610Z * [new branch] mlazos/tuple-fixes -> origin/mlazos/tuple-fixes 2025-09-07T07:36:00.2193981Z * [new branch] mlazos/tuple-fixes2 -> origin/mlazos/tuple-fixes2 2025-09-07T07:36:00.2194659Z * [new branch] mlazos/tuple-handling -> origin/mlazos/tuple-handling 2025-09-07T07:36:00.2195260Z * [new branch] mlazos/user-streams -> origin/mlazos/user-streams 2025-09-07T07:36:00.2195651Z * [new branch] mlazos/vary-beta -> origin/mlazos/vary-beta 2025-09-07T07:36:00.2196328Z * [new branch] mlazos/vary-beta2 -> origin/mlazos/vary-beta2 2025-09-07T07:36:00.2196930Z * [new branch] mlazos/weird-perf1 -> origin/mlazos/weird-perf1 2025-09-07T07:36:00.2197673Z * [new branch] mm_out_dtype_compile -> origin/mm_out_dtype_compile 2025-09-07T07:36:00.2198219Z * [new branch] modify-setupvllm -> origin/modify-setupvllm 2025-09-07T07:36:00.2198944Z * [new branch] module-shim -> origin/module-shim 2025-09-07T07:36:00.2199605Z * [new branch] move-theme-out-docker -> origin/move-theme-out-docker 2025-09-07T07:36:00.2200443Z * [new branch] msaroufim/be1 -> origin/msaroufim/be1 2025-09-07T07:36:00.2200909Z * [new branch] msaroufim/cn_path -> origin/msaroufim/cn_path 2025-09-07T07:36:00.2201573Z * [new branch] msaroufim/dtensorfusedadam -> origin/msaroufim/dtensorfusedadam 2025-09-07T07:36:00.2202127Z * [new branch] msaroufim/reduce -> origin/msaroufim/reduce 2025-09-07T07:36:00.2202951Z * [new branch] mtia/basic-cmake -> origin/mtia/basic-cmake 2025-09-07T07:36:00.2203486Z * [new branch] muon_dev -> origin/muon_dev 2025-09-07T07:36:00.2204124Z * [new branch] muon_dev_1 -> origin/muon_dev_1 2025-09-07T07:36:00.2204685Z * [new branch] nativert_num_outputs -> origin/nativert_num_outputs 2025-09-07T07:36:00.2205742Z * [new branch] nativert_numoutputs -> origin/nativert_numoutputs 2025-09-07T07:36:00.2206357Z * [new branch] new-modifiy-setupvllm -> origin/new-modifiy-setupvllm 2025-09-07T07:36:00.2206791Z * [new branch] new-setupvllm -> origin/new-setupvllm 2025-09-07T07:36:00.2207519Z * [new branch] new_zeros_dtype -> origin/new_zeros_dtype 2025-09-07T07:36:00.2208063Z * [new branch] newtest-base -> origin/newtest-base 2025-09-07T07:36:00.2209006Z * [new branch] ngimel/cat_perf1 -> origin/ngimel/cat_perf1 2025-09-07T07:36:00.2209393Z * [new branch] ngimel/einsum_fix -> origin/ngimel/einsum_fix 2025-09-07T07:36:00.2210004Z * [new branch] ngimel/error_index_list -> origin/ngimel/error_index_list 2025-09-07T07:36:00.2210398Z * [new branch] ngimel/fabric_check -> origin/ngimel/fabric_check 2025-09-07T07:36:00.2210936Z * [new branch] ngimel/fabric_fix -> origin/ngimel/fabric_fix 2025-09-07T07:36:00.2211400Z * [new branch] ngimel/fix_driver_init_error -> origin/ngimel/fix_driver_init_error 2025-09-07T07:36:00.2212186Z * [new branch] ngimel/fix_nccl_segment_seg -> origin/ngimel/fix_nccl_segment_seg 2025-09-07T07:36:00.2212829Z * [new branch] ngimel/gg_new -> origin/ngimel/gg_new 2025-09-07T07:36:00.2213529Z * [new branch] ngimel/modeguard -> origin/ngimel/modeguard 2025-09-07T07:36:00.2214207Z * [new branch] ngimel/multicast_fix -> origin/ngimel/multicast_fix 2025-09-07T07:36:00.2214838Z * [new branch] ngimel/rocm_handle_type -> origin/ngimel/rocm_handle_type 2025-09-07T07:36:00.2215395Z * [new branch] ngimel/symm_handle_fabric -> origin/ngimel/symm_handle_fabric 2025-09-07T07:36:00.2215852Z * [new branch] ngimel/unbind_multimem -> origin/ngimel/unbind_multimem 2025-09-07T07:36:00.2216532Z * [new branch] nightly -> origin/nightly 2025-09-07T07:36:00.2217373Z * [new branch] nmacchioni-patch-10 -> origin/nmacchioni-patch-10 2025-09-07T07:36:00.2218022Z * [new branch] nmacchioni-patch-7 -> origin/nmacchioni-patch-7 2025-09-07T07:36:00.2218696Z * [new branch] nmacchioni-patch-8 -> origin/nmacchioni-patch-8 2025-09-07T07:36:00.2219404Z * [new branch] nmacchioni-patch-9 -> origin/nmacchioni-patch-9 2025-09-07T07:36:00.2220109Z * [new branch] nullplay/fuse_matmul -> origin/nullplay/fuse_matmul 2025-09-07T07:36:00.2220707Z * [new branch] nullplay_fuse_matmul -> origin/nullplay_fuse_matmul 2025-09-07T07:36:00.2221345Z * [new branch] one-off -> origin/one-off 2025-09-07T07:36:00.2222307Z * [new branch] orig/release/1.10 -> origin/orig/release/1.10 2025-09-07T07:36:00.2222897Z * [new branch] orig/release/1.11 -> origin/orig/release/1.11 2025-09-07T07:36:00.2223701Z * [new branch] orig/release/1.12 -> origin/orig/release/1.12 2025-09-07T07:36:00.2224336Z * [new branch] orig/release/1.13 -> origin/orig/release/1.13 2025-09-07T07:36:00.2224937Z * [new branch] orig/release/1.6 -> origin/orig/release/1.6 2025-09-07T07:36:00.2225682Z * [new branch] orig/release/1.7 -> origin/orig/release/1.7 2025-09-07T07:36:00.2226715Z * [new branch] orig/release/1.8 -> origin/orig/release/1.8 2025-09-07T07:36:00.2227100Z * [new branch] orig/release/1.9 -> origin/orig/release/1.9 2025-09-07T07:36:00.2227806Z * [new branch] orig/release/2.0 -> origin/orig/release/2.0 2025-09-07T07:36:00.2228493Z * [new branch] orig/release/2.1 -> origin/orig/release/2.1 2025-09-07T07:36:00.2228924Z * [new branch] orig/release/2.2 -> origin/orig/release/2.2 2025-09-07T07:36:00.2229567Z * [new branch] orig/release/2.3 -> origin/orig/release/2.3 2025-09-07T07:36:00.2230157Z * [new branch] orig/release/2.4 -> origin/orig/release/2.4 2025-09-07T07:36:00.2230593Z * [new branch] orig/release/2.5 -> origin/orig/release/2.5 2025-09-07T07:36:00.2231396Z * [new branch] orig/release/2.6 -> origin/orig/release/2.6 2025-09-07T07:36:00.2232104Z * [new branch] orig/release/2.7 -> origin/orig/release/2.7 2025-09-07T07:36:00.2232738Z * [new branch] orig/release/2.8 -> origin/orig/release/2.8 2025-09-07T07:36:00.2233657Z * [new branch] oulgen/fx_graph -> origin/oulgen/fx_graph 2025-09-07T07:36:00.2234208Z * [new branch] padded-tensor -> origin/padded-tensor 2025-09-07T07:36:00.2234798Z * [new branch] pca2 -> origin/pca2 2025-09-07T07:36:00.2235439Z * [new branch] pianpwk-patch-1 -> origin/pianpwk-patch-1 2025-09-07T07:36:00.2236313Z * [new branch] pianpwk/backed_size_oblivious_export -> origin/pianpwk/backed_size_oblivious_export 2025-09-07T07:36:00.2236724Z * [new branch] pianpwk/invalidate_fake_memo -> origin/pianpwk/invalidate_fake_memo 2025-09-07T07:36:00.2237363Z * [new branch] pianpwk/max_1_strides -> origin/pianpwk/max_1_strides 2025-09-07T07:36:00.2237784Z * [new branch] pianpwk/maybe_guard_rel -> origin/pianpwk/maybe_guard_rel 2025-09-07T07:36:00.2238300Z * [new branch] pianpwk/nonzero_memo -> origin/pianpwk/nonzero_memo 2025-09-07T07:36:00.2239085Z * [new branch] pianpwk/oblivious_reshape_view_better -> origin/pianpwk/oblivious_reshape_view_better 2025-09-07T07:36:00.2239759Z * [new branch] pianpwk/oblivious_slice_forward -> origin/pianpwk/oblivious_slice_forward 2025-09-07T07:36:00.2240161Z * [new branch] pianpwk/oblivious_where -> origin/pianpwk/oblivious_where 2025-09-07T07:36:00.2240735Z * [new branch] pianpwk/param_static_pgo -> origin/pianpwk/param_static_pgo 2025-09-07T07:36:00.2241200Z * [new branch] pianpwk/pre_forward_hook -> origin/pianpwk/pre_forward_hook 2025-09-07T07:36:00.2241899Z * [new branch] pianpwk/remove_guard_fail_break -> origin/pianpwk/remove_guard_fail_break 2025-09-07T07:36:00.2242469Z * [new branch] pianpwk/slice_fresh_symbols -> origin/pianpwk/slice_fresh_symbols 2025-09-07T07:36:00.2242892Z * [new branch] pianpwk/sym_tokens_draft -> origin/pianpwk/sym_tokens_draft 2025-09-07T07:36:00.2243901Z * [new branch] pianpwk/test_pointwise_guard_or_false -> origin/pianpwk/test_pointwise_guard_or_false 2025-09-07T07:36:00.2244325Z * [new branch] pianpwk/test_slice_fake_impl -> origin/pianpwk/test_slice_fake_impl 2025-09-07T07:36:00.2244970Z * [new branch] pianpwk/totally_draft_sym_wrap -> origin/pianpwk/totally_draft_sym_wrap 2025-09-07T07:36:00.2245363Z * [new branch] pianpwk/unbacked_channels_last -> origin/pianpwk/unbacked_channels_last 2025-09-07T07:36:00.2246135Z * [new branch] pianpwk/unbacked_safe_conv1d -> origin/pianpwk/unbacked_safe_conv1d 2025-09-07T07:36:00.2246596Z * [new branch] pianpwk/unbacked_sdpa_flash -> origin/pianpwk/unbacked_sdpa_flash 2025-09-07T07:36:00.2247248Z * [new branch] pianpwk/unbacked_should_swap -> origin/pianpwk/unbacked_should_swap 2025-09-07T07:36:00.2247698Z * [new branch] pianpwk/unbacked_should_swap_2 -> origin/pianpwk/unbacked_should_swap_2 2025-09-07T07:36:00.2248270Z * [new branch] pianpwk/unbacked_slice_binding -> origin/pianpwk/unbacked_slice_binding 2025-09-07T07:36:00.2248737Z * [new branch] pianpwk/unbacked_slice_forward -> origin/pianpwk/unbacked_slice_forward 2025-09-07T07:36:00.2249372Z * [new branch] pianpwk/user_symints -> origin/pianpwk/user_symints 2025-09-07T07:36:00.2249798Z * [new branch] pianpwk/wan21_reshape -> origin/pianpwk/wan21_reshape 2025-09-07T07:36:00.2250375Z * [new branch] pianpwk/whitelist_optimizer -> origin/pianpwk/whitelist_optimizer 2025-09-07T07:36:00.2251076Z * [new branch] pin-torchao -> origin/pin-torchao 2025-09-07T07:36:00.2252005Z * [new branch] piz/fall_back_missing_0716 -> origin/piz/fall_back_missing_0716 2025-09-07T07:36:00.2252404Z * [new branch] piz/improve_scatter_0808 -> origin/piz/improve_scatter_0808 2025-09-07T07:36:00.2253101Z * [new branch] pool-separate -> origin/pool-separate 2025-09-07T07:36:00.2253641Z * [new branch] pr-156087 -> origin/pr-156087 2025-09-07T07:36:00.2254515Z * [new branch] pr/131860 -> origin/pr/131860 2025-09-07T07:36:00.2255223Z * [new branch] predispatch_to -> origin/predispatch_to 2025-09-07T07:36:00.2255840Z * [new branch] pt-opt-cuda3 -> origin/pt-opt-cuda3 2025-09-07T07:36:00.2256436Z * [new branch] pyobjectslot -> origin/pyobjectslot 2025-09-07T07:36:00.2257347Z * [new branch] python_compiled_autograd -> origin/python_compiled_autograd 2025-09-07T07:36:00.2258404Z * [new branch] qchip/export-D54134695 -> origin/qchip/export-D54134695 2025-09-07T07:36:00.2259005Z * [new branch] quint-bits -> origin/quint-bits 2025-09-07T07:36:00.2260207Z * [new branch] release/1.10 -> origin/release/1.10 2025-09-07T07:36:00.2260881Z * [new branch] release/1.11 -> origin/release/1.11 2025-09-07T07:36:00.2261563Z * [new branch] release/1.12 -> origin/release/1.12 2025-09-07T07:36:00.2262091Z * [new branch] release/1.13 -> origin/release/1.13 2025-09-07T07:36:00.2262682Z * [new branch] release/1.4 -> origin/release/1.4 2025-09-07T07:36:00.2262969Z * [new branch] release/1.4.1 -> origin/release/1.4.1 2025-09-07T07:36:00.2263642Z * [new branch] release/1.5 -> origin/release/1.5 2025-09-07T07:36:00.2264234Z * [new branch] release/1.6 -> origin/release/1.6 2025-09-07T07:36:00.2264848Z * [new branch] release/1.7 -> origin/release/1.7 2025-09-07T07:36:00.2265634Z * [new branch] release/1.8 -> origin/release/1.8 2025-09-07T07:36:00.2266207Z * [new branch] release/1.9 -> origin/release/1.9 2025-09-07T07:36:00.2266702Z * [new branch] release/2.0 -> origin/release/2.0 2025-09-07T07:36:00.2267479Z * [new branch] release/2.1 -> origin/release/2.1 2025-09-07T07:36:00.2268097Z * [new branch] release/2.2 -> origin/release/2.2 2025-09-07T07:36:00.2268942Z * [new branch] release/2.3 -> origin/release/2.3 2025-09-07T07:36:00.2269802Z * [new branch] release/2.4 -> origin/release/2.4 2025-09-07T07:36:00.2270738Z * [new branch] release/2.5 -> origin/release/2.5 2025-09-07T07:36:00.2271178Z * [new branch] release/2.6 -> origin/release/2.6 2025-09-07T07:36:00.2271868Z * [new branch] release/2.7 -> origin/release/2.7 2025-09-07T07:36:00.2272495Z * [new branch] release/2.8 -> origin/release/2.8 2025-09-07T07:36:00.2273144Z * [new branch] release_notes -> origin/release_notes 2025-09-07T07:36:00.2273772Z * [new branch] remove-actionable-label -> origin/remove-actionable-label 2025-09-07T07:36:00.2274360Z * [new branch] remove-ao -> origin/remove-ao 2025-09-07T07:36:00.2275085Z * [new branch] removedeprecatedvllmtest -> origin/removedeprecatedvllmtest 2025-09-07T07:36:00.2275869Z * [new branch] replace-pytorch-labs-20250812-195836 -> origin/replace-pytorch-labs-20250812-195836 2025-09-07T07:36:00.2276287Z * [new branch] replace-pytorch-labs-20250812-200248 -> origin/replace-pytorch-labs-20250812-200248 2025-09-07T07:36:00.2276768Z * [new branch] replace-pytorch-labs-20250812-200324 -> origin/replace-pytorch-labs-20250812-200324 2025-09-07T07:36:00.2277449Z * [new branch] replace-pytorch-labs-20250812-204020 -> origin/replace-pytorch-labs-20250812-204020 2025-09-07T07:36:00.2277893Z * [new branch] replace-pytorch-labs-20250812-204125 -> origin/replace-pytorch-labs-20250812-204125 2025-09-07T07:36:00.2278515Z * [new branch] replace-pytorch-labs-20250812-205624 -> origin/replace-pytorch-labs-20250812-205624 2025-09-07T07:36:00.2279834Z * [new branch] revert-131069-gh/krzysztofjordan/1/head -> origin/revert-131069-gh/krzysztofjordan/1/head 2025-09-07T07:36:00.2280986Z * [new branch] revert-131469-gh/andrewor14/51/head -> origin/revert-131469-gh/andrewor14/51/head 2025-09-07T07:36:00.2282224Z * [new branch] revert-156870-gh/skarjala/3/head -> origin/revert-156870-gh/skarjala/3/head 2025-09-07T07:36:00.2282926Z * [new branch] revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ -> origin/revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ 2025-09-07T07:36:00.2283295Z * [new branch] rocm-monitoring -> origin/rocm-monitoring 2025-09-07T07:36:00.2284133Z * [new branch] ruisi/relax_memory -> origin/ruisi/relax_memory 2025-09-07T07:36:00.2285221Z * [new branch] run-torchbench-smoke-test-h100 -> origin/run-torchbench-smoke-test-h100 2025-09-07T07:36:00.2286467Z * [new branch] ryanguo99/cleanup-dynamo-expected-failures -> origin/ryanguo99/cleanup-dynamo-expected-failures 2025-09-07T07:36:00.2286657Z * [new branch] ryanguo99/fix-closure-var -> origin/ryanguo99/fix-closure-var 2025-09-07T07:36:00.2286963Z * [new branch] rzou/faketensor_bench -> origin/rzou/faketensor_bench 2025-09-07T07:36:00.2287494Z * [new branch] rzou/njt -> origin/rzou/njt 2025-09-07T07:36:00.2288103Z * [new branch] rzou/pca -> origin/rzou/pca 2025-09-07T07:36:00.2288488Z * [new branch] rzou/realprop -> origin/rzou/realprop 2025-09-07T07:36:00.2289104Z * [new branch] rzou/setup_context -> origin/rzou/setup_context 2025-09-07T07:36:00.2290031Z * [new branch] sanchitintel/refactor_aten_int8_woq_gemm -> origin/sanchitintel/refactor_aten_int8_woq_gemm 2025-09-07T07:36:00.2290718Z * [new branch] sanchitintel/weird_thing_with_test_cpu_select_algorithm -> origin/sanchitintel/weird_thing_with_test_cpu_select_algorithm 2025-09-07T07:36:00.2291180Z * [new branch] sapling-pr-archive-SS-JIA -> origin/sapling-pr-archive-SS-JIA 2025-09-07T07:36:00.2291831Z * [new branch] save -> origin/save 2025-09-07T07:36:00.2292601Z * [new branch] sdym/2.5.1 -> origin/sdym/2.5.1 2025-09-07T07:36:00.2293242Z * [new branch] seemethere-patch-1 -> origin/seemethere-patch-1 2025-09-07T07:36:00.2293836Z * [new branch] setupvllm -> origin/setupvllm 2025-09-07T07:36:00.2294506Z * [new branch] share_and_pin_fork -> origin/share_and_pin_fork 2025-09-07T07:36:00.2295225Z * [new branch] shengf/fx-xform-perf -> origin/shengf/fx-xform-perf 2025-09-07T07:36:00.2295942Z * [new branch] shikaili_fp8_allgather -> origin/shikaili_fp8_allgather 2025-09-07T07:36:00.2296702Z * [new branch] shoumikhin-patch-1 -> origin/shoumikhin-patch-1 2025-09-07T07:36:00.2297254Z * [new branch] shoumikhin-patch-12 -> origin/shoumikhin-patch-12 2025-09-07T07:36:00.2297934Z * [new branch] simplify-fq-per-channel -> origin/simplify-fq-per-channel 2025-09-07T07:36:00.2298494Z * [new branch] solve-accuracy-fix -> origin/solve-accuracy-fix 2025-09-07T07:36:00.2299214Z * [new branch] soulitzer/stash-tls-ac -> origin/soulitzer/stash-tls-ac 2025-09-07T07:36:00.2300007Z * [new branch] sqzhang/flight4 -> origin/sqzhang/flight4 2025-09-07T07:36:00.2300644Z * [new branch] sqzhang/flight4plus -> origin/sqzhang/flight4plus 2025-09-07T07:36:00.2301350Z * [new branch] sraikund/record_funct_test -> origin/sraikund/record_funct_test 2025-09-07T07:36:00.2302184Z * [new branch] sraikund16/test -> origin/sraikund16/test 2025-09-07T07:36:00.2302830Z * [new branch] stablize-compilation-time -> origin/stablize-compilation-time 2025-09-07T07:36:00.2303448Z * [new branch] standalone-templates -> origin/standalone-templates 2025-09-07T07:36:00.2304137Z * [new branch] standalone_package_weights -> origin/standalone_package_weights 2025-09-07T07:36:00.2304698Z * [new branch] starterTaskUpdate -> origin/starterTaskUpdate 2025-09-07T07:36:00.2305374Z * [new branch] subgraph_fuse -> origin/subgraph_fuse 2025-09-07T07:36:00.2306098Z * [new branch] support-uv-in-collect_env -> origin/support-uv-in-collect_env 2025-09-07T07:36:00.2306962Z * [new branch] sve-poc -> origin/sve-poc 2025-09-07T07:36:00.2307592Z * [new branch] svekars-patch-1 -> origin/svekars-patch-1 2025-09-07T07:36:00.2308170Z * [new branch] switch-bn -> origin/switch-bn 2025-09-07T07:36:00.2308795Z * [new branch] sympy-bottleneck-repro -> origin/sympy-bottleneck-repro 2025-09-07T07:36:00.2309520Z * [new branch] tenpercent/ck_rocm_ci_v3 -> origin/tenpercent/ck_rocm_ci_v3 2025-09-07T07:36:00.2310169Z * [new branch] tensordict_integration -> origin/tensordict_integration 2025-09-07T07:36:00.2310739Z * [new branch] test-7054 -> origin/test-7054 2025-09-07T07:36:00.2311419Z * [new branch] test-move-conda-builds -> origin/test-move-conda-builds 2025-09-07T07:36:00.2312145Z * [new branch] test-myst-markdown-docstring -> origin/test-myst-markdown-docstring 2025-09-07T07:36:00.2312681Z * [new branch] test-old -> origin/test-old 2025-09-07T07:36:00.2313583Z * [new branch] test-vec-migration-internally -> origin/test-vec-migration-internally 2025-09-07T07:36:00.2314353Z * [new branch] test/bmm_heur -> origin/test/bmm_heur 2025-09-07T07:36:00.2314757Z * [new branch] test/inductor -> origin/test/inductor 2025-09-07T07:36:00.2315670Z * [new branch] tianren/flex_paged_attn_fix -> origin/tianren/flex_paged_attn_fix 2025-09-07T07:36:00.2316256Z * [new branch] tianren/flex_paged_attn_fix_temp -> origin/tianren/flex_paged_attn_fix_temp 2025-09-07T07:36:00.2316664Z * [new branch] tianren/test -> origin/tianren/test 2025-09-07T07:36:00.2317350Z * [new branch] tidy_performance_cyy -> origin/tidy_performance_cyy 2025-09-07T07:36:00.2317939Z * [new branch] torchtitan_ep -> origin/torchtitan_ep 2025-09-07T07:36:00.2318540Z * [new branch] trace_fsdp_torchtune_lora -> origin/trace_fsdp_torchtune_lora 2025-09-07T07:36:00.2319103Z * [new branch] traceable_fsdp_unit_tests -> origin/traceable_fsdp_unit_tests 2025-09-07T07:36:00.2319691Z * [new branch] tree_loop_vec_base -> origin/tree_loop_vec_base 2025-09-07T07:36:00.2320382Z * [new branch] tree_vec_base -> origin/tree_vec_base 2025-09-07T07:36:00.2321032Z * [new branch] triton-update -> origin/triton-update 2025-09-07T07:36:00.2321615Z * [new branch] triton_kernel -> origin/triton_kernel 2025-09-07T07:36:00.2322153Z * [new branch] triton_kernel_perf -> origin/triton_kernel_perf 2025-09-07T07:36:00.2322746Z * [new branch] tt_pkg_1908 -> origin/tt_pkg_1908 2025-09-07T07:36:00.2323417Z * [new branch] tweak-transformer-dependabot -> origin/tweak-transformer-dependabot 2025-09-07T07:36:00.2324042Z * [new branch] type_dec -> origin/type_dec 2025-09-07T07:36:00.2324666Z * [new branch] udate-sphinx-dependancies -> origin/udate-sphinx-dependancies 2025-09-07T07:36:00.2325598Z * [new branch] update-audio-commit-hash/16818882925-1712-1 -> origin/update-audio-commit-hash/16818882925-1712-1 2025-09-07T07:36:00.2326057Z * [new branch] update-audio-commit-hash/16895560422-1720-1 -> origin/update-audio-commit-hash/16895560422-1720-1 2025-09-07T07:36:00.2326588Z * [new branch] update-audio-commit-hash/16924174496-1738-1 -> origin/update-audio-commit-hash/16924174496-1738-1 2025-09-07T07:36:00.2327432Z * [new branch] update-audio-commit-hash/17002010821-1749-1 -> origin/update-audio-commit-hash/17002010821-1749-1 2025-09-07T07:36:00.2327864Z * [new branch] update-audio-commit-hash/17056004427-1766-1 -> origin/update-audio-commit-hash/17056004427-1766-1 2025-09-07T07:36:00.2328607Z * [new branch] update-audio-commit-hash/17085054029-1767-1 -> origin/update-audio-commit-hash/17085054029-1767-1 2025-09-07T07:36:00.2329280Z * [new branch] update-audio-commit-hash/17142507405-1771-1 -> origin/update-audio-commit-hash/17142507405-1771-1 2025-09-07T07:36:00.2329941Z * [new branch] update-audio-commit-hash/17168762740-1773-1 -> origin/update-audio-commit-hash/17168762740-1773-1 2025-09-07T07:36:00.2330568Z * [new branch] update-audio-commit-hash/17311174639-1780-1 -> origin/update-audio-commit-hash/17311174639-1780-1 2025-09-07T07:36:00.2331021Z * [new branch] update-audio-commit-hash/17336898740-1781-1 -> origin/update-audio-commit-hash/17336898740-1781-1 2025-09-07T07:36:00.2331544Z * [new branch] update-audio-commit-hash/17389727684-1786-1 -> origin/update-audio-commit-hash/17389727684-1786-1 2025-09-07T07:36:00.2332071Z * [new branch] update-audio-commit-hash/17449538142-1790-1 -> origin/update-audio-commit-hash/17449538142-1790-1 2025-09-07T07:36:00.2332691Z * [new branch] update-audio-commit-hash/17507351808-1794-1 -> origin/update-audio-commit-hash/17507351808-1794-1 2025-09-07T07:36:00.2333162Z * [new branch] update-dynamic-shapes-doc -> origin/update-dynamic-shapes-doc 2025-09-07T07:36:00.2334161Z * [new branch] update-executorch-commit-hash/15694981040-1626-1 -> origin/update-executorch-commit-hash/15694981040-1626-1 2025-09-07T07:36:00.2334908Z * [new branch] update-triton-commit-hash/13663274526-1487-2 -> origin/update-triton-commit-hash/13663274526-1487-2 2025-09-07T07:36:00.2335589Z * [new branch] update-vision-commit-hash/15336342773-1607-1 -> origin/update-vision-commit-hash/15336342773-1607-1 2025-09-07T07:36:00.2336529Z * [new branch] update-vllm-commit-hash/16737365217-1704-1 -> origin/update-vllm-commit-hash/16737365217-1704-1 2025-09-07T07:36:00.2336812Z * [new branch] update-vllm-commit-hash/16843157111-1713-1 -> origin/update-vllm-commit-hash/16843157111-1713-1 2025-09-07T07:36:00.2337430Z * [new branch] update-vllm-commit-hash/16855312394-1714-1 -> origin/update-vllm-commit-hash/16855312394-1714-1 2025-09-07T07:36:00.2337913Z * [new branch] update-vllm-commit-hash/16924174496-1738-1 -> origin/update-vllm-commit-hash/16924174496-1738-1 2025-09-07T07:36:00.2338417Z * [new branch] update-vllm-commit-hash/16952608705-1745-1 -> origin/update-vllm-commit-hash/16952608705-1745-1 2025-09-07T07:36:00.2339157Z * [new branch] update-vllm-commit-hash/16979836546-1748-1 -> origin/update-vllm-commit-hash/16979836546-1748-1 2025-09-07T07:36:00.2339782Z * [new branch] update-vllm-commit-hash/17014576881-1756-1 -> origin/update-vllm-commit-hash/17014576881-1756-1 2025-09-07T07:36:00.2340500Z * [new branch] update-vllm-commit-hash/17027830869-1761-1 -> origin/update-vllm-commit-hash/17027830869-1761-1 2025-09-07T07:36:00.2340990Z * [new branch] update-vllm-commit-hash/17056004427-1766-1 -> origin/update-vllm-commit-hash/17056004427-1766-1 2025-09-07T07:36:00.2341497Z * [new branch] update-vllm-commit-hash/17085054029-1767-1 -> origin/update-vllm-commit-hash/17085054029-1767-1 2025-09-07T07:36:00.2342195Z * [new branch] update-vllm-commit-hash/17113610216-1768-1 -> origin/update-vllm-commit-hash/17113610216-1768-1 2025-09-07T07:36:00.2342676Z * [new branch] update-vllm-commit-hash/17142507405-1771-1 -> origin/update-vllm-commit-hash/17142507405-1771-1 2025-09-07T07:36:00.2343363Z * [new branch] update-vllm-commit-hash/17181878974-1774-1 -> origin/update-vllm-commit-hash/17181878974-1774-1 2025-09-07T07:36:00.2343774Z * [new branch] update-vllm-commit-hash/17311174639-1780-1 -> origin/update-vllm-commit-hash/17311174639-1780-1 2025-09-07T07:36:00.2344398Z * [new branch] update-vllm-commit-hash/17336898740-1781-1 -> origin/update-vllm-commit-hash/17336898740-1781-1 2025-09-07T07:36:00.2344833Z * [new branch] update-vllm-commit-hash/17364352302-1785-1 -> origin/update-vllm-commit-hash/17364352302-1785-1 2025-09-07T07:36:00.2345449Z * [new branch] update-vllm-commit-hash/17389727684-1786-1 -> origin/update-vllm-commit-hash/17389727684-1786-1 2025-09-07T07:36:00.2345900Z * [new branch] update-vllm-commit-hash/17449538142-1790-1 -> origin/update-vllm-commit-hash/17449538142-1790-1 2025-09-07T07:36:00.2346409Z * [new branch] update-vllm-commit-hash/17480069797-1791-1 -> origin/update-vllm-commit-hash/17480069797-1791-1 2025-09-07T07:36:00.2347388Z * [new branch] update-vllm-commit-hash/17507351808-1794-1 -> origin/update-vllm-commit-hash/17507351808-1794-1 2025-09-07T07:36:00.2348108Z * [new branch] update-xla-commit-hash/16873912760-198-1 -> origin/update-xla-commit-hash/16873912760-198-1 2025-09-07T07:36:00.2348580Z * [new branch] update-xla-commit-hash/17034266655-199-1 -> origin/update-xla-commit-hash/17034266655-199-1 2025-09-07T07:36:00.2349045Z * [new branch] update-xla-commit-hash/17202464405-200-1 -> origin/update-xla-commit-hash/17202464405-200-1 2025-09-07T07:36:00.2349731Z * [new branch] update_docs_torch_multinomial_issue#125388 -> origin/update_docs_torch_multinomial_issue#125388 2025-09-07T07:36:00.2350196Z * [new branch] update_executorch_pin -> origin/update_executorch_pin 2025-09-07T07:36:00.2350920Z * [new branch] update_slow_tests_1722488736 -> origin/update_slow_tests_1722488736 2025-09-07T07:36:00.2351550Z * [new branch] update_slow_tests_1722879173 -> origin/update_slow_tests_1722879173 2025-09-07T07:36:00.2352191Z * [new branch] update_slow_tests_1752478971 -> origin/update_slow_tests_1752478971 2025-09-07T07:36:00.2352790Z * [new branch] update_slow_tests_1755502951 -> origin/update_slow_tests_1755502951 2025-09-07T07:36:00.2353349Z * [new branch] update_slow_tests_1756107664 -> origin/update_slow_tests_1756107664 2025-09-07T07:36:00.2354057Z * [new branch] update_submodule_FBGEMM -> origin/update_submodule_FBGEMM 2025-09-07T07:36:00.2354503Z * [new branch] update_submodule_kineto -> origin/update_submodule_kineto 2025-09-07T07:36:00.2355156Z * [new branch] update_submodule_tensorpipe -> origin/update_submodule_tensorpipe 2025-09-07T07:36:00.2355806Z * [new branch] v0.1.2 -> origin/v0.1.2 2025-09-07T07:36:00.2356548Z * [new branch] v1.0.1 -> origin/v1.0.1 2025-09-07T07:36:00.2357202Z * [new branch] v1.0.3 -> origin/v1.0.3 2025-09-07T07:36:00.2357818Z * [new branch] v1.1.0 -> origin/v1.1.0 2025-09-07T07:36:00.2358478Z * [new branch] v1.2.0 -> origin/v1.2.0 2025-09-07T07:36:00.2359104Z * [new branch] v1.3.0 -> origin/v1.3.0 2025-09-07T07:36:00.2359732Z * [new branch] v1.3.1 -> origin/v1.3.1 2025-09-07T07:36:00.2360505Z * [new branch] validate_fn -> origin/validate_fn 2025-09-07T07:36:00.2361240Z * [new branch] validations_2.6 -> origin/validations_2.6 2025-09-07T07:36:00.2362042Z * [new branch] validations_2.8 -> origin/validations_2.8 2025-09-07T07:36:00.2362886Z * [new branch] viable/strict -> origin/viable/strict 2025-09-07T07:36:00.2363424Z * [new branch] vllmbuildci -> origin/vllmbuildci 2025-09-07T07:36:00.2364076Z * [new branch] vllmpin -> origin/vllmpin 2025-09-07T07:36:00.2364879Z * [new branch] wdvr/conda_devcontainer -> origin/wdvr/conda_devcontainer 2025-09-07T07:36:00.2365295Z * [new branch] wdvr/iss_145259 -> origin/wdvr/iss_145259 2025-09-07T07:36:00.2366200Z * [new branch] weight_sharing_cpp -> origin/weight_sharing_cpp 2025-09-07T07:36:00.2367107Z * [new branch] whc/flight4 -> origin/whc/flight4 2025-09-07T07:36:00.2367638Z * [new branch] whc/flight51 -> origin/whc/flight51 2025-09-07T07:36:00.2368190Z * [new branch] whc/flight53 -> origin/whc/flight53 2025-09-07T07:36:00.2372697Z * [new branch] whc/stage2 -> origin/whc/stage2 2025-09-07T07:36:00.2373225Z * [new branch] whc/uneven -> origin/whc/uneven 2025-09-07T07:36:00.2374095Z * [new branch] whc/uneven-merge -> origin/whc/uneven-merge 2025-09-07T07:36:00.2374680Z * [new branch] win_warnings -> origin/win_warnings 2025-09-07T07:36:00.2375372Z * [new branch] windows_libtorch_free -> origin/windows_libtorch_free 2025-09-07T07:36:00.2375943Z * [new branch] workonoldcommit -> origin/workonoldcommit 2025-09-07T07:36:00.2376702Z * [new branch] wychi-autotune-prune-configs-by-shared-mem -> origin/wychi-autotune-prune-configs-by-shared-mem 2025-09-07T07:36:00.2377385Z * [new branch] xmfan/ca_0516 -> origin/xmfan/ca_0516 2025-09-07T07:36:00.2377854Z * [new branch] xmfan/ca_1051b93192 -> origin/xmfan/ca_1051b93192 2025-09-07T07:36:00.2378843Z * [new branch] xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 -> origin/xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 2025-09-07T07:36:00.2379037Z * [new branch] xmfan/ca_5a2be192d1 -> origin/xmfan/ca_5a2be192d1 2025-09-07T07:36:00.2379350Z * [new branch] xmfan/ca_9d59b516e9 -> origin/xmfan/ca_9d59b516e9 2025-09-07T07:36:00.2379961Z * [new branch] xmfan/ca_api -> origin/xmfan/ca_api 2025-09-07T07:36:00.2380385Z * [new branch] xmfan/ca_apr8 -> origin/xmfan/ca_apr8 2025-09-07T07:36:00.2381105Z * [new branch] xmfan/ca_base -> origin/xmfan/ca_base 2025-09-07T07:36:00.2381906Z * [new branch] xmfan/ca_cudagraphs -> origin/xmfan/ca_cudagraphs 2025-09-07T07:36:00.2382362Z * [new branch] xmfan/ca_dynamic -> origin/xmfan/ca_dynamic 2025-09-07T07:36:00.2382955Z * [new branch] xmfan/ca_fix_dyn -> origin/xmfan/ca_fix_dyn 2025-09-07T07:36:00.2383752Z * [new branch] xmfan/ca_fix_lowering -> origin/xmfan/ca_fix_lowering 2025-09-07T07:36:00.2384387Z * [new branch] xmfan/ca_fix_polyfills -> origin/xmfan/ca_fix_polyfills 2025-09-07T07:36:00.2384778Z * [new branch] xmfan/ca_jan3 -> origin/xmfan/ca_jan3 2025-09-07T07:36:00.2385359Z * [new branch] xmfan/ca_jun18 -> origin/xmfan/ca_jun18 2025-09-07T07:36:00.2385883Z * [new branch] xmfan/ca_jun24 -> origin/xmfan/ca_jun24 2025-09-07T07:36:00.2386419Z * [new branch] xmfan/ca_mem_base -> origin/xmfan/ca_mem_base 2025-09-07T07:36:00.2386848Z * [new branch] xmfan/ca_mem_fix -> origin/xmfan/ca_mem_fix 2025-09-07T07:36:00.2387473Z * [new branch] xmfan/ca_memory_fix -> origin/xmfan/ca_memory_fix 2025-09-07T07:36:00.2388099Z * [new branch] xmfan/ca_memory_fix_rebased -> origin/xmfan/ca_memory_fix_rebased 2025-09-07T07:36:00.2388803Z * [new branch] xmfan/ca_memory_fix_rebased2 -> origin/xmfan/ca_memory_fix_rebased2 2025-09-07T07:36:00.2389110Z * [new branch] xmfan/ca_move_to_cuda -> origin/xmfan/ca_move_to_cuda 2025-09-07T07:36:00.2389699Z * [new branch] xmfan/ca_nested -> origin/xmfan/ca_nested 2025-09-07T07:36:00.2390288Z * [new branch] xmfan/ca_overhead -> origin/xmfan/ca_overhead 2025-09-07T07:36:00.2390754Z * [new branch] xmfan/ca_overhead_0eba7e5451 -> origin/xmfan/ca_overhead_0eba7e5451 2025-09-07T07:36:00.2391310Z * [new branch] xmfan/ca_scalar -> origin/xmfan/ca_scalar 2025-09-07T07:36:00.2391895Z * [new branch] xmfan/ca_subclass_mem_fix -> origin/xmfan/ca_subclass_mem_fix 2025-09-07T07:36:00.2392414Z * [new branch] xmfan/ca_warm_mem -> origin/xmfan/ca_warm_mem 2025-09-07T07:36:00.2392919Z * [new branch] xmfan/ca_warm_mem_base -> origin/xmfan/ca_warm_mem_base 2025-09-07T07:36:00.2393520Z * [new branch] xmfan/cacu_jun18 -> origin/xmfan/cacu_jun18 2025-09-07T07:36:00.2393926Z * [new branch] xmfan/cacu_jun19 -> origin/xmfan/cacu_jun19 2025-09-07T07:36:00.2394532Z * [new branch] xmfan/cacu_jun4 -> origin/xmfan/cacu_jun4 2025-09-07T07:36:00.2395080Z * [new branch] xmfan/cacu_may27 -> origin/xmfan/cacu_may27 2025-09-07T07:36:00.2395652Z * [new branch] xmfan/disable_duck_shape -> origin/xmfan/disable_duck_shape 2025-09-07T07:36:00.2396105Z * [new branch] xmfan/fca_cpp_node_passthrough -> origin/xmfan/fca_cpp_node_passthrough 2025-09-07T07:36:00.2396720Z * [new branch] xmfan/issue_123374 -> origin/xmfan/issue_123374 2025-09-07T07:36:00.2397682Z * [new branch] xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 2025-09-07T07:36:00.2398140Z * [new branch] xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 2025-09-07T07:36:00.2398752Z * [new branch] xmfan/segfault_test -> origin/xmfan/segfault_test 2025-09-07T07:36:00.2399160Z * [new branch] xmfan/single_step -> origin/xmfan/single_step 2025-09-07T07:36:00.2399780Z * [new branch] xmfan/sth_0829 -> origin/xmfan/sth_0829 2025-09-07T07:36:00.2400392Z * [new branch] xmfan/test -> origin/xmfan/test 2025-09-07T07:36:00.2401333Z * [new branch] yguo/debug-0226-constexpr -> origin/yguo/debug-0226-constexpr 2025-09-07T07:36:00.2401772Z * [new branch] yguo/new_latest_changes -> origin/yguo/new_latest_changes 2025-09-07T07:36:00.2402401Z * [new branch] yguo/patch_constexpr_changes -> origin/yguo/patch_constexpr_changes 2025-09-07T07:36:00.2402958Z * [new branch] yihan_quantization -> origin/yihan_quantization 2025-09-07T07:36:00.2403785Z * [new branch] yiming/add_jit_trace_benchmark -> origin/yiming/add_jit_trace_benchmark 2025-09-07T07:36:00.2404175Z * [new branch] yiming/add_nativert_benchmark -> origin/yiming/add_nativert_benchmark 2025-09-07T07:36:00.2404716Z * [new branch] yiming/bootcamp -> origin/yiming/bootcamp 2025-09-07T07:36:00.2405596Z * [new branch] zainr/canary-test -> origin/zainr/canary-test 2025-09-07T07:36:00.2406170Z * [new branch] zainr/cleanup-gh-runners -> origin/zainr/cleanup-gh-runners 2025-09-07T07:36:00.2406579Z * [new branch] zainr/git-push-v2 -> origin/zainr/git-push-v2 2025-09-07T07:36:00.2407189Z * [new branch] zainr/pull-migration-c -> origin/zainr/pull-migration-c 2025-09-07T07:36:00.2407606Z * [new branch] zainr/test -> origin/zainr/test 2025-09-07T07:36:00.2408239Z * [new branch] zainr/test2 -> origin/zainr/test2 2025-09-07T07:36:00.2408610Z * [new branch] zainr/unstable -> origin/zainr/unstable 2025-09-07T07:36:00.2409221Z * [new branch] zainr/unstable-xla -> origin/zainr/unstable-xla 2025-09-07T07:36:00.2409933Z * [new branch] zasdfgbnm-patch-3 -> origin/zasdfgbnm-patch-3 2025-09-07T07:36:00.2410564Z * [new branch] zb2p -> origin/zb2p 2025-09-07T07:36:00.2411173Z * [new branch] zero_grad_optimization -> origin/zero_grad_optimization 2025-09-07T07:36:00.2411803Z * [new branch] zeros-and-scatter-part2 -> origin/zeros-and-scatter-part2 2025-09-07T07:36:00.2412646Z * [new branch] zhxchen17/scratch/0 -> origin/zhxchen17/scratch/0 2025-09-07T07:36:00.2413517Z * [new branch] zhxhcen17/moodycamel -> origin/zhxhcen17/moodycamel 2025-09-07T07:36:00.2414275Z * [new branch] zxiiro/main -> origin/zxiiro/main 2025-09-07T07:36:00.2415006Z * [new tag] bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug -> bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug 2025-09-07T07:36:00.2415304Z * [new tag] ci/binaries/77164 -> ci/binaries/77164 2025-09-07T07:36:00.2415890Z * [new tag] ciflow/binaries/156049 -> ciflow/binaries/156049 2025-09-07T07:36:00.2416184Z * [new tag] ciflow/binaries/156712 -> ciflow/binaries/156712 2025-09-07T07:36:00.2416496Z * [new tag] ciflow/binaries/157432 -> ciflow/binaries/157432 2025-09-07T07:36:00.2416899Z * [new tag] ciflow/binaries/157685 -> ciflow/binaries/157685 2025-09-07T07:36:00.2417287Z * [new tag] ciflow/binaries/157689 -> ciflow/binaries/157689 2025-09-07T07:36:00.2417570Z * [new tag] ciflow/binaries/158104 -> ciflow/binaries/158104 2025-09-07T07:36:00.2417993Z * [new tag] ciflow/binaries/160229 -> ciflow/binaries/160229 2025-09-07T07:36:00.2418356Z * [new tag] ciflow/binaries/160720 -> ciflow/binaries/160720 2025-09-07T07:36:00.2418751Z * [new tag] ciflow/binaries/162080 -> ciflow/binaries/162080 2025-09-07T07:36:00.2419071Z * [new tag] ciflow/binaries/162329 -> ciflow/binaries/162329 2025-09-07T07:36:00.2419505Z * [new tag] ciflow/binaries_libtorch/156049 -> ciflow/binaries_libtorch/156049 2025-09-07T07:36:00.2419817Z * [new tag] ciflow/binaries_libtorch/156711 -> ciflow/binaries_libtorch/156711 2025-09-07T07:36:00.2420272Z * [new tag] ciflow/binaries_libtorch/157432 -> ciflow/binaries_libtorch/157432 2025-09-07T07:36:00.2420636Z * [new tag] ciflow/binaries_wheel/156049 -> ciflow/binaries_wheel/156049 2025-09-07T07:36:00.2421007Z * [new tag] ciflow/binaries_wheel/156711 -> ciflow/binaries_wheel/156711 2025-09-07T07:36:00.2421288Z * [new tag] ciflow/binaries_wheel/157432 -> ciflow/binaries_wheel/157432 2025-09-07T07:36:00.2421618Z * [new tag] ciflow/binaries_wheel/162136 -> ciflow/binaries_wheel/162136 2025-09-07T07:36:00.2422063Z * [new tag] ciflow/binaries_wheel/162252 -> ciflow/binaries_wheel/162252 2025-09-07T07:36:00.2422423Z * [new tag] ciflow/binaries_wheel/162325 -> ciflow/binaries_wheel/162325 2025-09-07T07:36:00.2422987Z * [new tag] ciflow/h100-distributed/156703 -> ciflow/h100-distributed/156703 2025-09-07T07:36:00.2423297Z * [new tag] ciflow/h100-symm-mem/157635 -> ciflow/h100-symm-mem/157635 2025-09-07T07:36:00.2423637Z * [new tag] ciflow/h100-symm-mem/161984 -> ciflow/h100-symm-mem/161984 2025-09-07T07:36:00.2423968Z * [new tag] ciflow/h100-symm-mem/162003 -> ciflow/h100-symm-mem/162003 2025-09-07T07:36:00.2424336Z * [new tag] ciflow/h100-symm-mem/162011 -> ciflow/h100-symm-mem/162011 2025-09-07T07:36:00.2424583Z * [new tag] ciflow/h100-symm-mem/162026 -> ciflow/h100-symm-mem/162026 2025-09-07T07:36:00.2424939Z * [new tag] ciflow/h100-symm-mem/162033 -> ciflow/h100-symm-mem/162033 2025-09-07T07:36:00.2425249Z * [new tag] ciflow/h100-symm-mem/162040 -> ciflow/h100-symm-mem/162040 2025-09-07T07:36:00.2425581Z * [new tag] ciflow/h100-symm-mem/162041 -> ciflow/h100-symm-mem/162041 2025-09-07T07:36:00.2425881Z * [new tag] ciflow/h100-symm-mem/162142 -> ciflow/h100-symm-mem/162142 2025-09-07T07:36:00.2426258Z * [new tag] ciflow/h100-symm-mem/162150 -> ciflow/h100-symm-mem/162150 2025-09-07T07:36:00.2426580Z * [new tag] ciflow/h100-symm-mem/162243 -> ciflow/h100-symm-mem/162243 2025-09-07T07:36:00.2427188Z * [new tag] ciflow/h100-symm-mem/162320 -> ciflow/h100-symm-mem/162320 2025-09-07T07:36:00.2427541Z * [new tag] ciflow/h100/159158 -> ciflow/h100/159158 2025-09-07T07:36:00.2428323Z * [new tag] ciflow/h100/160480 -> ciflow/h100/160480 2025-09-07T07:36:00.2429003Z * [new tag] ciflow/h100/161749 -> ciflow/h100/161749 2025-09-07T07:36:00.2429290Z * [new tag] ciflow/h100/162022 -> ciflow/h100/162022 2025-09-07T07:36:00.2429639Z * [new tag] ciflow/h100/162278 -> ciflow/h100/162278 2025-09-07T07:36:00.2430355Z * [new tag] ciflow/inductor-perf-test-nightly-rocm/156592 -> ciflow/inductor-perf-test-nightly-rocm/156592 2025-09-07T07:36:00.2430790Z * [new tag] ciflow/inductor-perf-test-nightly/156592 -> ciflow/inductor-perf-test-nightly/156592 2025-09-07T07:36:00.2431178Z * [new tag] ciflow/inductor-periodic/162063 -> ciflow/inductor-periodic/162063 2025-09-07T07:36:00.2431527Z * [new tag] ciflow/inductor-periodic/162227 -> ciflow/inductor-periodic/162227 2025-09-07T07:36:00.2432113Z * [new tag] ciflow/inductor-periodic/162323 -> ciflow/inductor-periodic/162323 2025-09-07T07:36:00.2432661Z * [new tag] ciflow/inductor-rocm/154170 -> ciflow/inductor-rocm/154170 2025-09-07T07:36:00.2433075Z * [new tag] ciflow/inductor-rocm/159146 -> ciflow/inductor-rocm/159146 2025-09-07T07:36:00.2433366Z * [new tag] ciflow/inductor-rocm/159158 -> ciflow/inductor-rocm/159158 2025-09-07T07:36:00.2433922Z * [new tag] ciflow/inductor-rocm/161715 -> ciflow/inductor-rocm/161715 2025-09-07T07:36:00.2434253Z * [new tag] ciflow/inductor-rocm/162053 -> ciflow/inductor-rocm/162053 2025-09-07T07:36:00.2435077Z * [new tag] ciflow/inductor-rocm/162056 -> ciflow/inductor-rocm/162056 2025-09-07T07:36:00.2435414Z * [new tag] ciflow/inductor/137400 -> ciflow/inductor/137400 2025-09-07T07:36:00.2435723Z * [new tag] ciflow/inductor/148180 -> ciflow/inductor/148180 2025-09-07T07:36:00.2436058Z * [new tag] ciflow/inductor/148328 -> ciflow/inductor/148328 2025-09-07T07:36:00.2436367Z * [new tag] ciflow/inductor/148484 -> ciflow/inductor/148484 2025-09-07T07:36:00.2436730Z * [new tag] ciflow/inductor/148492 -> ciflow/inductor/148492 2025-09-07T07:36:00.2437019Z * [new tag] ciflow/inductor/152624 -> ciflow/inductor/152624 2025-09-07T07:36:00.2437375Z * [new tag] ciflow/inductor/154694 -> ciflow/inductor/154694 2025-09-07T07:36:00.2437688Z * [new tag] ciflow/inductor/156049 -> ciflow/inductor/156049 2025-09-07T07:36:00.2438012Z * [new tag] ciflow/inductor/156592 -> ciflow/inductor/156592 2025-09-07T07:36:00.2438357Z * [new tag] ciflow/inductor/157635 -> ciflow/inductor/157635 2025-09-07T07:36:00.2438711Z * [new tag] ciflow/inductor/157685 -> ciflow/inductor/157685 2025-09-07T07:36:00.2439329Z * [new tag] ciflow/inductor/157686 -> ciflow/inductor/157686 2025-09-07T07:36:00.2439657Z * [new tag] ciflow/inductor/157689 -> ciflow/inductor/157689 2025-09-07T07:36:00.2440172Z * [new tag] ciflow/inductor/157699 -> ciflow/inductor/157699 2025-09-07T07:36:00.2440601Z * [new tag] ciflow/inductor/157743 -> ciflow/inductor/157743 2025-09-07T07:36:00.2440985Z * [new tag] ciflow/inductor/157994 -> ciflow/inductor/157994 2025-09-07T07:36:00.2441315Z * [new tag] ciflow/inductor/158091 -> ciflow/inductor/158091 2025-09-07T07:36:00.2441670Z * [new tag] ciflow/inductor/158104 -> ciflow/inductor/158104 2025-09-07T07:36:00.2442462Z * [new tag] ciflow/inductor/158404 -> ciflow/inductor/158404 2025-09-07T07:36:00.2442779Z * [new tag] ciflow/inductor/158647 -> ciflow/inductor/158647 2025-09-07T07:36:00.2443206Z * [new tag] ciflow/inductor/158932 -> ciflow/inductor/158932 2025-09-07T07:36:00.2443530Z * [new tag] ciflow/inductor/159146 -> ciflow/inductor/159146 2025-09-07T07:36:00.2443881Z * [new tag] ciflow/inductor/159158 -> ciflow/inductor/159158 2025-09-07T07:36:00.2444538Z * [new tag] ciflow/inductor/159274 -> ciflow/inductor/159274 2025-09-07T07:36:00.2444843Z * [new tag] ciflow/inductor/159664 -> ciflow/inductor/159664 2025-09-07T07:36:00.2445420Z * [new tag] ciflow/inductor/159778 -> ciflow/inductor/159778 2025-09-07T07:36:00.2445655Z * [new tag] ciflow/inductor/159835 -> ciflow/inductor/159835 2025-09-07T07:36:00.2446132Z * [new tag] ciflow/inductor/159944 -> ciflow/inductor/159944 2025-09-07T07:36:00.2446686Z * [new tag] ciflow/inductor/160161 -> ciflow/inductor/160161 2025-09-07T07:36:00.2446883Z * [new tag] ciflow/inductor/160174 -> ciflow/inductor/160174 2025-09-07T07:36:00.2447514Z * [new tag] ciflow/inductor/160323 -> ciflow/inductor/160323 2025-09-07T07:36:00.2447940Z * [new tag] ciflow/inductor/160324 -> ciflow/inductor/160324 2025-09-07T07:36:00.2448539Z * [new tag] ciflow/inductor/160325 -> ciflow/inductor/160325 2025-09-07T07:36:00.2448889Z * [new tag] ciflow/inductor/160326 -> ciflow/inductor/160326 2025-09-07T07:36:00.2449294Z * [new tag] ciflow/inductor/160327 -> ciflow/inductor/160327 2025-09-07T07:36:00.2449655Z * [new tag] ciflow/inductor/160328 -> ciflow/inductor/160328 2025-09-07T07:36:00.2450183Z * [new tag] ciflow/inductor/160329 -> ciflow/inductor/160329 2025-09-07T07:36:00.2450450Z * [new tag] ciflow/inductor/160480 -> ciflow/inductor/160480 2025-09-07T07:36:00.2450984Z * [new tag] ciflow/inductor/160532 -> ciflow/inductor/160532 2025-09-07T07:36:00.2451824Z * [new tag] ciflow/inductor/160539 -> ciflow/inductor/160539 2025-09-07T07:36:00.2452136Z * [new tag] ciflow/inductor/160580 -> ciflow/inductor/160580 2025-09-07T07:36:00.2452457Z * [new tag] ciflow/inductor/160685 -> ciflow/inductor/160685 2025-09-07T07:36:00.2452775Z * [new tag] ciflow/inductor/160686 -> ciflow/inductor/160686 2025-09-07T07:36:00.2453116Z * [new tag] ciflow/inductor/160687 -> ciflow/inductor/160687 2025-09-07T07:36:00.2453494Z * [new tag] ciflow/inductor/160688 -> ciflow/inductor/160688 2025-09-07T07:36:00.2453839Z * [new tag] ciflow/inductor/160690 -> ciflow/inductor/160690 2025-09-07T07:36:00.2454159Z * [new tag] ciflow/inductor/160706 -> ciflow/inductor/160706 2025-09-07T07:36:00.2454562Z * [new tag] ciflow/inductor/160729 -> ciflow/inductor/160729 2025-09-07T07:36:00.2454957Z * [new tag] ciflow/inductor/160798 -> ciflow/inductor/160798 2025-09-07T07:36:00.2455388Z * [new tag] ciflow/inductor/160836 -> ciflow/inductor/160836 2025-09-07T07:36:00.2455747Z * [new tag] ciflow/inductor/160843 -> ciflow/inductor/160843 2025-09-07T07:36:00.2456387Z * [new tag] ciflow/inductor/160869 -> ciflow/inductor/160869 2025-09-07T07:36:00.2456763Z * [new tag] ciflow/inductor/160920 -> ciflow/inductor/160920 2025-09-07T07:36:00.2457115Z * [new tag] ciflow/inductor/160943 -> ciflow/inductor/160943 2025-09-07T07:36:00.2457444Z * [new tag] ciflow/inductor/161092 -> ciflow/inductor/161092 2025-09-07T07:36:00.2457811Z * [new tag] ciflow/inductor/161093 -> ciflow/inductor/161093 2025-09-07T07:36:00.2458360Z * [new tag] ciflow/inductor/161109 -> ciflow/inductor/161109 2025-09-07T07:36:00.2458672Z * [new tag] ciflow/inductor/161118 -> ciflow/inductor/161118 2025-09-07T07:36:00.2459276Z * [new tag] ciflow/inductor/161178 -> ciflow/inductor/161178 2025-09-07T07:36:00.2459534Z * [new tag] ciflow/inductor/161246 -> ciflow/inductor/161246 2025-09-07T07:36:00.2459880Z * [new tag] ciflow/inductor/161349 -> ciflow/inductor/161349 2025-09-07T07:36:00.2460234Z * [new tag] ciflow/inductor/161350 -> ciflow/inductor/161350 2025-09-07T07:36:00.2460570Z * [new tag] ciflow/inductor/161351 -> ciflow/inductor/161351 2025-09-07T07:36:00.2461001Z * [new tag] ciflow/inductor/161397 -> ciflow/inductor/161397 2025-09-07T07:36:00.2461340Z * [new tag] ciflow/inductor/161404 -> ciflow/inductor/161404 2025-09-07T07:36:00.2461678Z * [new tag] ciflow/inductor/161405 -> ciflow/inductor/161405 2025-09-07T07:36:00.2462034Z * [new tag] ciflow/inductor/161406 -> ciflow/inductor/161406 2025-09-07T07:36:00.2462608Z * [new tag] ciflow/inductor/161410 -> ciflow/inductor/161410 2025-09-07T07:36:00.2462891Z * [new tag] ciflow/inductor/161414 -> ciflow/inductor/161414 2025-09-07T07:36:00.2463500Z * [new tag] ciflow/inductor/161442 -> ciflow/inductor/161442 2025-09-07T07:36:00.2463813Z * [new tag] ciflow/inductor/161458 -> ciflow/inductor/161458 2025-09-07T07:36:00.2464160Z * [new tag] ciflow/inductor/161468 -> ciflow/inductor/161468 2025-09-07T07:36:00.2464460Z * [new tag] ciflow/inductor/161469 -> ciflow/inductor/161469 2025-09-07T07:36:00.2465008Z * [new tag] ciflow/inductor/161485 -> ciflow/inductor/161485 2025-09-07T07:36:00.2465267Z * [new tag] ciflow/inductor/161499 -> ciflow/inductor/161499 2025-09-07T07:36:00.2465588Z * [new tag] ciflow/inductor/161534 -> ciflow/inductor/161534 2025-09-07T07:36:00.2465940Z * [new tag] ciflow/inductor/161595 -> ciflow/inductor/161595 2025-09-07T07:36:00.2466289Z * [new tag] ciflow/inductor/161596 -> ciflow/inductor/161596 2025-09-07T07:36:00.2467046Z * [new tag] ciflow/inductor/161630 -> ciflow/inductor/161630 2025-09-07T07:36:00.2467319Z * [new tag] ciflow/inductor/161667 -> ciflow/inductor/161667 2025-09-07T07:36:00.2467631Z * [new tag] ciflow/inductor/161670 -> ciflow/inductor/161670 2025-09-07T07:36:00.2468206Z * [new tag] ciflow/inductor/161673 -> ciflow/inductor/161673 2025-09-07T07:36:00.2468869Z * [new tag] ciflow/inductor/161674 -> ciflow/inductor/161674 2025-09-07T07:36:00.2469236Z * [new tag] ciflow/inductor/161675 -> ciflow/inductor/161675 2025-09-07T07:36:00.2469500Z * [new tag] ciflow/inductor/161693 -> ciflow/inductor/161693 2025-09-07T07:36:00.2469800Z * [new tag] ciflow/inductor/161695 -> ciflow/inductor/161695 2025-09-07T07:36:00.2470192Z * [new tag] ciflow/inductor/161715 -> ciflow/inductor/161715 2025-09-07T07:36:00.2470540Z * [new tag] ciflow/inductor/161730 -> ciflow/inductor/161730 2025-09-07T07:36:00.2470858Z * [new tag] ciflow/inductor/161732 -> ciflow/inductor/161732 2025-09-07T07:36:00.2471444Z * [new tag] ciflow/inductor/161744 -> ciflow/inductor/161744 2025-09-07T07:36:00.2471717Z * [new tag] ciflow/inductor/161746 -> ciflow/inductor/161746 2025-09-07T07:36:00.2472094Z * [new tag] ciflow/inductor/161747 -> ciflow/inductor/161747 2025-09-07T07:36:00.2472446Z * [new tag] ciflow/inductor/161819 -> ciflow/inductor/161819 2025-09-07T07:36:00.2472764Z * [new tag] ciflow/inductor/161821 -> ciflow/inductor/161821 2025-09-07T07:36:00.2473083Z * [new tag] ciflow/inductor/161828 -> ciflow/inductor/161828 2025-09-07T07:36:00.2473476Z * [new tag] ciflow/inductor/161879 -> ciflow/inductor/161879 2025-09-07T07:36:00.2473775Z * [new tag] ciflow/inductor/161880 -> ciflow/inductor/161880 2025-09-07T07:36:00.2474152Z * [new tag] ciflow/inductor/161881 -> ciflow/inductor/161881 2025-09-07T07:36:00.2474679Z * [new tag] ciflow/inductor/161907 -> ciflow/inductor/161907 2025-09-07T07:36:00.2474952Z * [new tag] ciflow/inductor/161914 -> ciflow/inductor/161914 2025-09-07T07:36:00.2475387Z * [new tag] ciflow/inductor/161924 -> ciflow/inductor/161924 2025-09-07T07:36:00.2475798Z * [new tag] ciflow/inductor/161936 -> ciflow/inductor/161936 2025-09-07T07:36:00.2476167Z * [new tag] ciflow/inductor/161938 -> ciflow/inductor/161938 2025-09-07T07:36:00.2476556Z * [new tag] ciflow/inductor/161939 -> ciflow/inductor/161939 2025-09-07T07:36:00.2476901Z * [new tag] ciflow/inductor/161940 -> ciflow/inductor/161940 2025-09-07T07:36:00.2477230Z * [new tag] ciflow/inductor/161955 -> ciflow/inductor/161955 2025-09-07T07:36:00.2477546Z * [new tag] ciflow/inductor/161957 -> ciflow/inductor/161957 2025-09-07T07:36:00.2477936Z * [new tag] ciflow/inductor/161975 -> ciflow/inductor/161975 2025-09-07T07:36:00.2478323Z * [new tag] ciflow/inductor/161977 -> ciflow/inductor/161977 2025-09-07T07:36:00.2478650Z * [new tag] ciflow/inductor/161978 -> ciflow/inductor/161978 2025-09-07T07:36:00.2479199Z * [new tag] ciflow/inductor/161979 -> ciflow/inductor/161979 2025-09-07T07:36:00.2479471Z * [new tag] ciflow/inductor/161980 -> ciflow/inductor/161980 2025-09-07T07:36:00.2479811Z * [new tag] ciflow/inductor/161988 -> ciflow/inductor/161988 2025-09-07T07:36:00.2480134Z * [new tag] ciflow/inductor/161994 -> ciflow/inductor/161994 2025-09-07T07:36:00.2480494Z * [new tag] ciflow/inductor/162013 -> ciflow/inductor/162013 2025-09-07T07:36:00.2480810Z * [new tag] ciflow/inductor/162014 -> ciflow/inductor/162014 2025-09-07T07:36:00.2481185Z * [new tag] ciflow/inductor/162017 -> ciflow/inductor/162017 2025-09-07T07:36:00.2481571Z * [new tag] ciflow/inductor/162021 -> ciflow/inductor/162021 2025-09-07T07:36:00.2481892Z * [new tag] ciflow/inductor/162023 -> ciflow/inductor/162023 2025-09-07T07:36:00.2482207Z * [new tag] ciflow/inductor/162027 -> ciflow/inductor/162027 2025-09-07T07:36:00.2482610Z * [new tag] ciflow/inductor/162029 -> ciflow/inductor/162029 2025-09-07T07:36:00.2482988Z * [new tag] ciflow/inductor/162030 -> ciflow/inductor/162030 2025-09-07T07:36:00.2483341Z * [new tag] ciflow/inductor/162031 -> ciflow/inductor/162031 2025-09-07T07:36:00.2483694Z * [new tag] ciflow/inductor/162033 -> ciflow/inductor/162033 2025-09-07T07:36:00.2484258Z * [new tag] ciflow/inductor/162052 -> ciflow/inductor/162052 2025-09-07T07:36:00.2484483Z * [new tag] ciflow/inductor/162053 -> ciflow/inductor/162053 2025-09-07T07:36:00.2484841Z * [new tag] ciflow/inductor/162056 -> ciflow/inductor/162056 2025-09-07T07:36:00.2485229Z * [new tag] ciflow/inductor/162063 -> ciflow/inductor/162063 2025-09-07T07:36:00.2485552Z * [new tag] ciflow/inductor/162066 -> ciflow/inductor/162066 2025-09-07T07:36:00.2485952Z * [new tag] ciflow/inductor/162068 -> ciflow/inductor/162068 2025-09-07T07:36:00.2486521Z * [new tag] ciflow/inductor/162081 -> ciflow/inductor/162081 2025-09-07T07:36:00.2486807Z * [new tag] ciflow/inductor/162088 -> ciflow/inductor/162088 2025-09-07T07:36:00.2487140Z * [new tag] ciflow/inductor/162089 -> ciflow/inductor/162089 2025-09-07T07:36:00.2487489Z * [new tag] ciflow/inductor/162094 -> ciflow/inductor/162094 2025-09-07T07:36:00.2487843Z * [new tag] ciflow/inductor/162098 -> ciflow/inductor/162098 2025-09-07T07:36:00.2488152Z * [new tag] ciflow/inductor/162101 -> ciflow/inductor/162101 2025-09-07T07:36:00.2488536Z * [new tag] ciflow/inductor/162102 -> ciflow/inductor/162102 2025-09-07T07:36:00.2488882Z * [new tag] ciflow/inductor/162104 -> ciflow/inductor/162104 2025-09-07T07:36:00.2489218Z * [new tag] ciflow/inductor/162106 -> ciflow/inductor/162106 2025-09-07T07:36:00.2489570Z * [new tag] ciflow/inductor/162108 -> ciflow/inductor/162108 2025-09-07T07:36:00.2490030Z * [new tag] ciflow/inductor/162126 -> ciflow/inductor/162126 2025-09-07T07:36:00.2490408Z * [new tag] ciflow/inductor/162149 -> ciflow/inductor/162149 2025-09-07T07:36:00.2490740Z * [new tag] ciflow/inductor/162164 -> ciflow/inductor/162164 2025-09-07T07:36:00.2491092Z * [new tag] ciflow/inductor/162166 -> ciflow/inductor/162166 2025-09-07T07:36:00.2491437Z * [new tag] ciflow/inductor/162169 -> ciflow/inductor/162169 2025-09-07T07:36:00.2491782Z * [new tag] ciflow/inductor/162170 -> ciflow/inductor/162170 2025-09-07T07:36:00.2492494Z * [new tag] ciflow/inductor/162171 -> ciflow/inductor/162171 2025-09-07T07:36:00.2492748Z * [new tag] ciflow/inductor/162183 -> ciflow/inductor/162183 2025-09-07T07:36:00.2493102Z * [new tag] ciflow/inductor/162189 -> ciflow/inductor/162189 2025-09-07T07:36:00.2493454Z * [new tag] ciflow/inductor/162190 -> ciflow/inductor/162190 2025-09-07T07:36:00.2493857Z * [new tag] ciflow/inductor/162191 -> ciflow/inductor/162191 2025-09-07T07:36:00.2494201Z * [new tag] ciflow/inductor/162194 -> ciflow/inductor/162194 2025-09-07T07:36:00.2494737Z * [new tag] ciflow/inductor/162200 -> ciflow/inductor/162200 2025-09-07T07:36:00.2495072Z * [new tag] ciflow/inductor/162201 -> ciflow/inductor/162201 2025-09-07T07:36:00.2495420Z * [new tag] ciflow/inductor/162208 -> ciflow/inductor/162208 2025-09-07T07:36:00.2495952Z * [new tag] ciflow/inductor/162211 -> ciflow/inductor/162211 2025-09-07T07:36:00.2496299Z * [new tag] ciflow/inductor/162216 -> ciflow/inductor/162216 2025-09-07T07:36:00.2496587Z * [new tag] ciflow/inductor/162220 -> ciflow/inductor/162220 2025-09-07T07:36:00.2497171Z * [new tag] ciflow/inductor/162222 -> ciflow/inductor/162222 2025-09-07T07:36:00.2497363Z * [new tag] ciflow/inductor/162227 -> ciflow/inductor/162227 2025-09-07T07:36:00.2497714Z * [new tag] ciflow/inductor/162238 -> ciflow/inductor/162238 2025-09-07T07:36:00.2498090Z * [new tag] ciflow/inductor/162239 -> ciflow/inductor/162239 2025-09-07T07:36:00.2498455Z * [new tag] ciflow/inductor/162240 -> ciflow/inductor/162240 2025-09-07T07:36:00.2498759Z * [new tag] ciflow/inductor/162244 -> ciflow/inductor/162244 2025-09-07T07:36:00.2499106Z * [new tag] ciflow/inductor/162245 -> ciflow/inductor/162245 2025-09-07T07:36:00.2499452Z * [new tag] ciflow/inductor/162262 -> ciflow/inductor/162262 2025-09-07T07:36:00.2499839Z * [new tag] ciflow/inductor/162275 -> ciflow/inductor/162275 2025-09-07T07:36:00.2500153Z * [new tag] ciflow/inductor/162278 -> ciflow/inductor/162278 2025-09-07T07:36:00.2500519Z * [new tag] ciflow/inductor/162284 -> ciflow/inductor/162284 2025-09-07T07:36:00.2501071Z * [new tag] ciflow/inductor/162286 -> ciflow/inductor/162286 2025-09-07T07:36:00.2501293Z * [new tag] ciflow/inductor/162288 -> ciflow/inductor/162288 2025-09-07T07:36:00.2501676Z * [new tag] ciflow/inductor/162293 -> ciflow/inductor/162293 2025-09-07T07:36:00.2502008Z * [new tag] ciflow/inductor/162294 -> ciflow/inductor/162294 2025-09-07T07:36:00.2502387Z * [new tag] ciflow/inductor/162295 -> ciflow/inductor/162295 2025-09-07T07:36:00.2502701Z * [new tag] ciflow/inductor/162296 -> ciflow/inductor/162296 2025-09-07T07:36:00.2503052Z * [new tag] ciflow/inductor/162298 -> ciflow/inductor/162298 2025-09-07T07:36:00.2503452Z * [new tag] ciflow/inductor/162307 -> ciflow/inductor/162307 2025-09-07T07:36:00.2503856Z * [new tag] ciflow/inductor/162309 -> ciflow/inductor/162309 2025-09-07T07:36:00.2504169Z * [new tag] ciflow/inductor/162311 -> ciflow/inductor/162311 2025-09-07T07:36:00.2504504Z * [new tag] ciflow/inductor/162312 -> ciflow/inductor/162312 2025-09-07T07:36:00.2504888Z * [new tag] ciflow/inductor/162315 -> ciflow/inductor/162315 2025-09-07T07:36:00.2505244Z * [new tag] ciflow/inductor/162316 -> ciflow/inductor/162316 2025-09-07T07:36:00.2505599Z * [new tag] ciflow/inductor/162318 -> ciflow/inductor/162318 2025-09-07T07:36:00.2505939Z * [new tag] ciflow/inductor/162323 -> ciflow/inductor/162323 2025-09-07T07:36:00.2506355Z * [new tag] ciflow/inductor/162341 -> ciflow/inductor/162341 2025-09-07T07:36:00.2506752Z * [new tag] ciflow/inductor/162345 -> ciflow/inductor/162345 2025-09-07T07:36:00.2507350Z * [new tag] ciflow/inductor/3b9a386 -> ciflow/inductor/3b9a386 2025-09-07T07:36:00.2507771Z * [new tag] ciflow/inductor/3d4b92b -> ciflow/inductor/3d4b92b 2025-09-07T07:36:00.2508279Z * [new tag] ciflow/inductor/d224ac7 -> ciflow/inductor/d224ac7 2025-09-07T07:36:00.2508702Z * [new tag] ciflow/linux-aarch64/157994 -> ciflow/linux-aarch64/157994 2025-09-07T07:36:00.2509015Z * [new tag] ciflow/linux-aarch64/159737 -> ciflow/linux-aarch64/159737 2025-09-07T07:36:00.2509363Z * [new tag] ciflow/linux-aarch64/160078 -> ciflow/linux-aarch64/160078 2025-09-07T07:36:00.2509767Z * [new tag] ciflow/mps/157553 -> ciflow/mps/157553 2025-09-07T07:36:00.2510115Z * [new tag] ciflow/mps/157635 -> ciflow/mps/157635 2025-09-07T07:36:00.2510443Z * [new tag] ciflow/mps/161988 -> ciflow/mps/161988 2025-09-07T07:36:00.2510777Z * [new tag] ciflow/mps/162108 -> ciflow/mps/162108 2025-09-07T07:36:00.2511077Z * [new tag] ciflow/mps/162153 -> ciflow/mps/162153 2025-09-07T07:36:00.2511410Z * [new tag] ciflow/mps/162281 -> ciflow/mps/162281 2025-09-07T07:36:00.2512039Z * [new tag] ciflow/nightly/156049 -> ciflow/nightly/156049 2025-09-07T07:36:00.2512348Z * [new tag] ciflow/nightly/158104 -> ciflow/nightly/158104 2025-09-07T07:36:00.2512754Z * [new tag] ciflow/op-benchmark/157994 -> ciflow/op-benchmark/157994 2025-09-07T07:36:00.2513401Z * [new tag] ciflow/periodic-rocm-mi300/161529 -> ciflow/periodic-rocm-mi300/161529 2025-09-07T07:36:00.2513628Z * [new tag] ciflow/periodic-rocm-mi300/161715 -> ciflow/periodic-rocm-mi300/161715 2025-09-07T07:36:00.2514240Z * [new tag] ciflow/periodic/054a2fd -> ciflow/periodic/054a2fd 2025-09-07T07:36:00.2514458Z * [new tag] ciflow/periodic/156703 -> ciflow/periodic/156703 2025-09-07T07:36:00.2514824Z * [new tag] ciflow/periodic/161715 -> ciflow/periodic/161715 2025-09-07T07:36:00.2515178Z * [new tag] ciflow/periodic/162021 -> ciflow/periodic/162021 2025-09-07T07:36:00.2515447Z * [new tag] ciflow/periodic/162323 -> ciflow/periodic/162323 2025-09-07T07:36:00.2516025Z * [new tag] ciflow/periodic/2a6d37d -> ciflow/periodic/2a6d37d 2025-09-07T07:36:00.2516372Z * [new tag] ciflow/periodic/317eeb8 -> ciflow/periodic/317eeb8 2025-09-07T07:36:00.2516783Z * [new tag] ciflow/periodic/3c32 -> ciflow/periodic/3c32 2025-09-07T07:36:00.2517384Z * [new tag] ciflow/periodic/3e98831 -> ciflow/periodic/3e98831 2025-09-07T07:36:00.2517829Z * [new tag] ciflow/periodic/94512-point -> ciflow/periodic/94512-point 2025-09-07T07:36:00.2518440Z * [new tag] ciflow/periodic/csl/test87519 -> ciflow/periodic/csl/test87519 2025-09-07T07:36:00.2519116Z * [new tag] ciflow/periodic/csltest88275 -> ciflow/periodic/csltest88275 2025-09-07T07:36:00.2519555Z * [new tag] ciflow/periodic/csltest88761 -> ciflow/periodic/csltest88761 2025-09-07T07:36:00.2520143Z * [new tag] ciflow/periodic/release_1.12 -> ciflow/periodic/release_1.12 2025-09-07T07:36:00.2520992Z * [new tag] ciflow/periodic/release_1.12.0 -> ciflow/periodic/release_1.12.0 2025-09-07T07:36:00.2521543Z * [new tag] ciflow/periodic/sha-ec5b83 -> ciflow/periodic/sha-ec5b83 2025-09-07T07:36:00.2521864Z * [new tag] ciflow/rocm-mi300/154170 -> ciflow/rocm-mi300/154170 2025-09-07T07:36:00.2522275Z * [new tag] ciflow/rocm-mi300/158747 -> ciflow/rocm-mi300/158747 2025-09-07T07:36:00.2522664Z * [new tag] ciflow/rocm-mi300/159146 -> ciflow/rocm-mi300/159146 2025-09-07T07:36:00.2522968Z * [new tag] ciflow/rocm-mi300/159158 -> ciflow/rocm-mi300/159158 2025-09-07T07:36:00.2523283Z * [new tag] ciflow/rocm-mi300/161715 -> ciflow/rocm-mi300/161715 2025-09-07T07:36:00.2523610Z * [new tag] ciflow/rocm-mi300/161957 -> ciflow/rocm-mi300/161957 2025-09-07T07:36:00.2523928Z * [new tag] ciflow/rocm-mi300/162053 -> ciflow/rocm-mi300/162053 2025-09-07T07:36:00.2524289Z * [new tag] ciflow/rocm-mi300/162056 -> ciflow/rocm-mi300/162056 2025-09-07T07:36:00.2524897Z * [new tag] ciflow/rocm-mi300/162112 -> ciflow/rocm-mi300/162112 2025-09-07T07:36:00.2525131Z * [new tag] ciflow/rocm-mi300/162245 -> ciflow/rocm-mi300/162245 2025-09-07T07:36:00.2525536Z * [new tag] ciflow/rocm-mi300/162278 -> ciflow/rocm-mi300/162278 2025-09-07T07:36:00.2525954Z * [new tag] ciflow/rocm-mi300/162288 -> ciflow/rocm-mi300/162288 2025-09-07T07:36:00.2526464Z * [new tag] ciflow/rocm-mi355/162053 -> ciflow/rocm-mi355/162053 2025-09-07T07:36:00.2526694Z * [new tag] ciflow/rocm-mi355/162056 -> ciflow/rocm-mi355/162056 2025-09-07T07:36:00.2527218Z * [new tag] ciflow/rocm/148492 -> ciflow/rocm/148492 2025-09-07T07:36:00.2527429Z * [new tag] ciflow/rocm/154170 -> ciflow/rocm/154170 2025-09-07T07:36:00.2528007Z * [new tag] ciflow/rocm/156491 -> ciflow/rocm/156491 2025-09-07T07:36:00.2528269Z * [new tag] ciflow/rocm/156592 -> ciflow/rocm/156592 2025-09-07T07:36:00.2528588Z * [new tag] ciflow/rocm/158747 -> ciflow/rocm/158747 2025-09-07T07:36:00.2528918Z * [new tag] ciflow/rocm/159146 -> ciflow/rocm/159146 2025-09-07T07:36:00.2529534Z * [new tag] ciflow/rocm/159158 -> ciflow/rocm/159158 2025-09-07T07:36:00.2529726Z * [new tag] ciflow/rocm/161715 -> ciflow/rocm/161715 2025-09-07T07:36:00.2530288Z * [new tag] ciflow/rocm/161972 -> ciflow/rocm/161972 2025-09-07T07:36:00.2530506Z * [new tag] ciflow/rocm/162052 -> ciflow/rocm/162052 2025-09-07T07:36:00.2530831Z * [new tag] ciflow/rocm/162053 -> ciflow/rocm/162053 2025-09-07T07:36:00.2531356Z * [new tag] ciflow/rocm/162056 -> ciflow/rocm/162056 2025-09-07T07:36:00.2531762Z * [new tag] ciflow/rocm/162112 -> ciflow/rocm/162112 2025-09-07T07:36:00.2532352Z * [new tag] ciflow/rocm/162278 -> ciflow/rocm/162278 2025-09-07T07:36:00.2532547Z * [new tag] ciflow/rocm/162288 -> ciflow/rocm/162288 2025-09-07T07:36:00.2532921Z * [new tag] ciflow/rocm/162305 -> ciflow/rocm/162305 2025-09-07T07:36:00.2533478Z * [new tag] ciflow/slow/01c7106 -> ciflow/slow/01c7106 2025-09-07T07:36:00.2533837Z * [new tag] ciflow/slow/0577043 -> ciflow/slow/0577043 2025-09-07T07:36:00.2534595Z * [new tag] ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym -> ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym 2025-09-07T07:36:00.2534786Z * [new tag] ciflow/slow/0e81104 -> ciflow/slow/0e81104 2025-09-07T07:36:00.2535138Z * [new tag] ciflow/slow/161395 -> ciflow/slow/161395 2025-09-07T07:36:00.2535590Z * [new tag] ciflow/slow/1732077 -> ciflow/slow/1732077 2025-09-07T07:36:00.2536126Z * [new tag] ciflow/slow/187eb7c -> ciflow/slow/187eb7c 2025-09-07T07:36:00.2536667Z * [new tag] ciflow/slow/1faef89 -> ciflow/slow/1faef89 2025-09-07T07:36:00.2537221Z * [new tag] ciflow/slow/3920ec1 -> ciflow/slow/3920ec1 2025-09-07T07:36:00.2537699Z * [new tag] ciflow/slow/3b7c6b2 -> ciflow/slow/3b7c6b2 2025-09-07T07:36:00.2538163Z * [new tag] ciflow/slow/59a3759 -> ciflow/slow/59a3759 2025-09-07T07:36:00.2538569Z * [new tag] ciflow/slow/70ef0bb -> ciflow/slow/70ef0bb 2025-09-07T07:36:00.2539038Z * [new tag] ciflow/slow/788ff06 -> ciflow/slow/788ff06 2025-09-07T07:36:00.2539766Z * [new tag] ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym -> ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym 2025-09-07T07:36:00.2539984Z * [new tag] ciflow/slow/9d85864 -> ciflow/slow/9d85864 2025-09-07T07:36:00.2540468Z * [new tag] ciflow/slow/9ffad5b -> ciflow/slow/9ffad5b 2025-09-07T07:36:00.2541070Z * [new tag] ciflow/slow/a206e8b -> ciflow/slow/a206e8b 2025-09-07T07:36:00.2541346Z * [new tag] ciflow/slow/a837609 -> ciflow/slow/a837609 2025-09-07T07:36:00.2541843Z * [new tag] ciflow/slow/af841f3 -> ciflow/slow/af841f3 2025-09-07T07:36:00.2542558Z * [new tag] ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym -> ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym 2025-09-07T07:36:00.2542919Z * [new tag] ciflow/triton_binaries/162329 -> ciflow/triton_binaries/162329 2025-09-07T07:36:00.2543272Z * [new tag] ciflow/trunk/113258 -> ciflow/trunk/113258 2025-09-07T07:36:00.2543648Z * [new tag] ciflow/trunk/137400 -> ciflow/trunk/137400 2025-09-07T07:36:00.2543918Z * [new tag] ciflow/trunk/148180 -> ciflow/trunk/148180 2025-09-07T07:36:00.2544251Z * [new tag] ciflow/trunk/148328 -> ciflow/trunk/148328 2025-09-07T07:36:00.2544592Z * [new tag] ciflow/trunk/148492 -> ciflow/trunk/148492 2025-09-07T07:36:00.2545183Z * [new tag] ciflow/trunk/148919 -> ciflow/trunk/148919 2025-09-07T07:36:00.2545417Z * [new tag] ciflow/trunk/152624 -> ciflow/trunk/152624 2025-09-07T07:36:00.2545698Z * [new tag] ciflow/trunk/154170 -> ciflow/trunk/154170 2025-09-07T07:36:00.2546053Z * [new tag] ciflow/trunk/154694 -> ciflow/trunk/154694 2025-09-07T07:36:00.2546383Z * [new tag] ciflow/trunk/156049 -> ciflow/trunk/156049 2025-09-07T07:36:00.2546723Z * [new tag] ciflow/trunk/156703 -> ciflow/trunk/156703 2025-09-07T07:36:00.2547145Z * [new tag] ciflow/trunk/156711 -> ciflow/trunk/156711 2025-09-07T07:36:00.2547610Z * [new tag] ciflow/trunk/157432 -> ciflow/trunk/157432 2025-09-07T07:36:00.2548020Z * [new tag] ciflow/trunk/157685 -> ciflow/trunk/157685 2025-09-07T07:36:00.2548399Z * [new tag] ciflow/trunk/157689 -> ciflow/trunk/157689 2025-09-07T07:36:00.2548685Z * [new tag] ciflow/trunk/157699 -> ciflow/trunk/157699 2025-09-07T07:36:00.2549277Z * [new tag] ciflow/trunk/157813 -> ciflow/trunk/157813 2025-09-07T07:36:00.2549508Z * [new tag] ciflow/trunk/157994 -> ciflow/trunk/157994 2025-09-07T07:36:00.2549902Z * [new tag] ciflow/trunk/158091 -> ciflow/trunk/158091 2025-09-07T07:36:00.2550197Z * [new tag] ciflow/trunk/158104 -> ciflow/trunk/158104 2025-09-07T07:36:00.2550588Z * [new tag] ciflow/trunk/158404 -> ciflow/trunk/158404 2025-09-07T07:36:00.2550941Z * [new tag] ciflow/trunk/158647 -> ciflow/trunk/158647 2025-09-07T07:36:00.2551522Z * [new tag] ciflow/trunk/158846 -> ciflow/trunk/158846 2025-09-07T07:36:00.2551820Z * [new tag] ciflow/trunk/159158 -> ciflow/trunk/159158 2025-09-07T07:36:00.2552208Z * [new tag] ciflow/trunk/159682 -> ciflow/trunk/159682 2025-09-07T07:36:00.2552555Z * [new tag] ciflow/trunk/159835 -> ciflow/trunk/159835 2025-09-07T07:36:00.2552882Z * [new tag] ciflow/trunk/160161 -> ciflow/trunk/160161 2025-09-07T07:36:00.2553231Z * [new tag] ciflow/trunk/160236 -> ciflow/trunk/160236 2025-09-07T07:36:00.2553566Z * [new tag] ciflow/trunk/160329 -> ciflow/trunk/160329 2025-09-07T07:36:00.2553956Z * [new tag] ciflow/trunk/160480 -> ciflow/trunk/160480 2025-09-07T07:36:00.2554358Z * [new tag] ciflow/trunk/160532 -> ciflow/trunk/160532 2025-09-07T07:36:00.2554733Z * [new tag] ciflow/trunk/160836 -> ciflow/trunk/160836 2025-09-07T07:36:00.2555138Z * [new tag] ciflow/trunk/160843 -> ciflow/trunk/160843 2025-09-07T07:36:00.2555436Z * [new tag] ciflow/trunk/160869 -> ciflow/trunk/160869 2025-09-07T07:36:00.2555968Z * [new tag] ciflow/trunk/160940 -> ciflow/trunk/160940 2025-09-07T07:36:00.2556174Z * [new tag] ciflow/trunk/160943 -> ciflow/trunk/160943 2025-09-07T07:36:00.2556742Z * [new tag] ciflow/trunk/160953 -> ciflow/trunk/160953 2025-09-07T07:36:00.2557101Z * [new tag] ciflow/trunk/161035 -> ciflow/trunk/161035 2025-09-07T07:36:00.2557488Z * [new tag] ciflow/trunk/161178 -> ciflow/trunk/161178 2025-09-07T07:36:00.2557758Z * [new tag] ciflow/trunk/161349 -> ciflow/trunk/161349 2025-09-07T07:36:00.2558134Z * [new tag] ciflow/trunk/161350 -> ciflow/trunk/161350 2025-09-07T07:36:00.2558485Z * [new tag] ciflow/trunk/161351 -> ciflow/trunk/161351 2025-09-07T07:36:00.2558848Z * [new tag] ciflow/trunk/161395 -> ciflow/trunk/161395 2025-09-07T07:36:00.2559174Z * [new tag] ciflow/trunk/161405 -> ciflow/trunk/161405 2025-09-07T07:36:00.2559484Z * [new tag] ciflow/trunk/161406 -> ciflow/trunk/161406 2025-09-07T07:36:00.2560022Z * [new tag] ciflow/trunk/161410 -> ciflow/trunk/161410 2025-09-07T07:36:00.2560263Z * [new tag] ciflow/trunk/161468 -> ciflow/trunk/161468 2025-09-07T07:36:00.2560618Z * [new tag] ciflow/trunk/161499 -> ciflow/trunk/161499 2025-09-07T07:36:00.2561285Z * [new tag] ciflow/trunk/161527 -> ciflow/trunk/161527 2025-09-07T07:36:00.2561483Z * [new tag] ciflow/trunk/161534 -> ciflow/trunk/161534 2025-09-07T07:36:00.2561819Z * [new tag] ciflow/trunk/161591 -> ciflow/trunk/161591 2025-09-07T07:36:00.2562185Z * [new tag] ciflow/trunk/161595 -> ciflow/trunk/161595 2025-09-07T07:36:00.2562535Z * [new tag] ciflow/trunk/161596 -> ciflow/trunk/161596 2025-09-07T07:36:00.2562894Z * [new tag] ciflow/trunk/161633 -> ciflow/trunk/161633 2025-09-07T07:36:00.2563230Z * [new tag] ciflow/trunk/161634 -> ciflow/trunk/161634 2025-09-07T07:36:00.2563579Z * [new tag] ciflow/trunk/161635 -> ciflow/trunk/161635 2025-09-07T07:36:00.2563913Z * [new tag] ciflow/trunk/161667 -> ciflow/trunk/161667 2025-09-07T07:36:00.2564202Z * [new tag] ciflow/trunk/161670 -> ciflow/trunk/161670 2025-09-07T07:36:00.2564593Z * [new tag] ciflow/trunk/161692 -> ciflow/trunk/161692 2025-09-07T07:36:00.2564902Z * [new tag] ciflow/trunk/161693 -> ciflow/trunk/161693 2025-09-07T07:36:00.2565286Z * [new tag] ciflow/trunk/161695 -> ciflow/trunk/161695 2025-09-07T07:36:00.2565593Z * [new tag] ciflow/trunk/161730 -> ciflow/trunk/161730 2025-09-07T07:36:00.2566393Z * [new tag] ciflow/trunk/161744 -> ciflow/trunk/161744 2025-09-07T07:36:00.2566654Z * [new tag] ciflow/trunk/161749 -> ciflow/trunk/161749 2025-09-07T07:36:00.2567005Z * [new tag] ciflow/trunk/161881 -> ciflow/trunk/161881 2025-09-07T07:36:00.2567338Z * [new tag] ciflow/trunk/161924 -> ciflow/trunk/161924 2025-09-07T07:36:00.2567933Z * [new tag] ciflow/trunk/161926 -> ciflow/trunk/161926 2025-09-07T07:36:00.2568185Z * [new tag] ciflow/trunk/161936 -> ciflow/trunk/161936 2025-09-07T07:36:00.2568545Z * [new tag] ciflow/trunk/161952 -> ciflow/trunk/161952 2025-09-07T07:36:00.2568992Z * [new tag] ciflow/trunk/161955 -> ciflow/trunk/161955 2025-09-07T07:36:00.2569253Z * [new tag] ciflow/trunk/161957 -> ciflow/trunk/161957 2025-09-07T07:36:00.2569631Z * [new tag] ciflow/trunk/161959 -> ciflow/trunk/161959 2025-09-07T07:36:00.2569935Z * [new tag] ciflow/trunk/161977 -> ciflow/trunk/161977 2025-09-07T07:36:00.2570266Z * [new tag] ciflow/trunk/161988 -> ciflow/trunk/161988 2025-09-07T07:36:00.2570840Z * [new tag] ciflow/trunk/161994 -> ciflow/trunk/161994 2025-09-07T07:36:00.2571207Z * [new tag] ciflow/trunk/162007 -> ciflow/trunk/162007 2025-09-07T07:36:00.2571527Z * [new tag] ciflow/trunk/162013 -> ciflow/trunk/162013 2025-09-07T07:36:00.2571883Z * [new tag] ciflow/trunk/162017 -> ciflow/trunk/162017 2025-09-07T07:36:00.2572231Z * [new tag] ciflow/trunk/162021 -> ciflow/trunk/162021 2025-09-07T07:36:00.2572536Z * [new tag] ciflow/trunk/162022 -> ciflow/trunk/162022 2025-09-07T07:36:00.2572919Z * [new tag] ciflow/trunk/162040 -> ciflow/trunk/162040 2025-09-07T07:36:00.2573212Z * [new tag] ciflow/trunk/162041 -> ciflow/trunk/162041 2025-09-07T07:36:00.2573738Z * [new tag] ciflow/trunk/162062 -> ciflow/trunk/162062 2025-09-07T07:36:00.2573971Z * [new tag] ciflow/trunk/162066 -> ciflow/trunk/162066 2025-09-07T07:36:00.2574369Z * [new tag] ciflow/trunk/162089 -> ciflow/trunk/162089 2025-09-07T07:36:00.2574782Z * [new tag] ciflow/trunk/162099 -> ciflow/trunk/162099 2025-09-07T07:36:00.2575117Z * [new tag] ciflow/trunk/162104 -> ciflow/trunk/162104 2025-09-07T07:36:00.2575473Z * [new tag] ciflow/trunk/162106 -> ciflow/trunk/162106 2025-09-07T07:36:00.2575798Z * [new tag] ciflow/trunk/162112 -> ciflow/trunk/162112 2025-09-07T07:36:00.2576165Z * [new tag] ciflow/trunk/162119 -> ciflow/trunk/162119 2025-09-07T07:36:00.2576536Z * [new tag] ciflow/trunk/162142 -> ciflow/trunk/162142 2025-09-07T07:36:00.2576910Z * [new tag] ciflow/trunk/162169 -> ciflow/trunk/162169 2025-09-07T07:36:00.2577210Z * [new tag] ciflow/trunk/162183 -> ciflow/trunk/162183 2025-09-07T07:36:00.2577567Z * [new tag] ciflow/trunk/162190 -> ciflow/trunk/162190 2025-09-07T07:36:00.2577971Z * [new tag] ciflow/trunk/162194 -> ciflow/trunk/162194 2025-09-07T07:36:00.2578350Z * [new tag] ciflow/trunk/162200 -> ciflow/trunk/162200 2025-09-07T07:36:00.2578726Z * [new tag] ciflow/trunk/162206 -> ciflow/trunk/162206 2025-09-07T07:36:00.2579040Z * [new tag] ciflow/trunk/162208 -> ciflow/trunk/162208 2025-09-07T07:36:00.2579423Z * [new tag] ciflow/trunk/162222 -> ciflow/trunk/162222 2025-09-07T07:36:00.2579766Z * [new tag] ciflow/trunk/162238 -> ciflow/trunk/162238 2025-09-07T07:36:00.2580088Z * [new tag] ciflow/trunk/162244 -> ciflow/trunk/162244 2025-09-07T07:36:00.2580768Z * [new tag] ciflow/trunk/162267 -> ciflow/trunk/162267 2025-09-07T07:36:00.2580994Z * [new tag] ciflow/trunk/162269 -> ciflow/trunk/162269 2025-09-07T07:36:00.2581529Z * [new tag] ciflow/trunk/162278 -> ciflow/trunk/162278 2025-09-07T07:36:00.2581815Z * [new tag] ciflow/trunk/162286 -> ciflow/trunk/162286 2025-09-07T07:36:00.2582106Z * [new tag] ciflow/trunk/162288 -> ciflow/trunk/162288 2025-09-07T07:36:00.2582483Z * [new tag] ciflow/trunk/162293 -> ciflow/trunk/162293 2025-09-07T07:36:00.2582896Z * [new tag] ciflow/trunk/162310 -> ciflow/trunk/162310 2025-09-07T07:36:00.2583187Z * [new tag] ciflow/trunk/162311 -> ciflow/trunk/162311 2025-09-07T07:36:00.2583503Z * [new tag] ciflow/trunk/162315 -> ciflow/trunk/162315 2025-09-07T07:36:00.2583858Z * [new tag] ciflow/trunk/162325 -> ciflow/trunk/162325 2025-09-07T07:36:00.2584380Z * [new tag] ciflow/trunk/162328 -> ciflow/trunk/162328 2025-09-07T07:36:00.2584621Z * [new tag] ciflow/trunk/162329 -> ciflow/trunk/162329 2025-09-07T07:36:00.2585318Z * [new tag] ciflow/unstable/123 -> ciflow/unstable/123 2025-09-07T07:36:00.2585674Z * [new tag] ciflow/vllm/162292 -> ciflow/vllm/162292 2025-09-07T07:36:00.2586118Z * [new tag] ciflow/win-arm64/156049 -> ciflow/win-arm64/156049 2025-09-07T07:36:00.2586444Z * [new tag] ciflow/win-arm64/158104 -> ciflow/win-arm64/158104 2025-09-07T07:36:00.2587218Z * [new tag] ciflow/xpu/157699 -> ciflow/xpu/157699 2025-09-07T07:36:00.2587447Z * [new tag] ciflow/xpu/157994 -> ciflow/xpu/157994 2025-09-07T07:36:00.2587858Z * [new tag] ciflow/xpu/159459 -> ciflow/xpu/159459 2025-09-07T07:36:00.2588157Z * [new tag] ciflow/xpu/159718 -> ciflow/xpu/159718 2025-09-07T07:36:00.2588501Z * [new tag] ciflow/xpu/159944 -> ciflow/xpu/159944 2025-09-07T07:36:00.2589068Z * [new tag] ciflow/xpu/160867 -> ciflow/xpu/160867 2025-09-07T07:36:00.2589368Z * [new tag] ciflow/xpu/160938 -> ciflow/xpu/160938 2025-09-07T07:36:00.2589684Z * [new tag] ciflow/xpu/160940 -> ciflow/xpu/160940 2025-09-07T07:36:00.2589992Z * [new tag] ciflow/xpu/160953 -> ciflow/xpu/160953 2025-09-07T07:36:00.2590527Z * [new tag] ciflow/xpu/161045 -> ciflow/xpu/161045 2025-09-07T07:36:00.2590917Z * [new tag] ciflow/xpu/161058 -> ciflow/xpu/161058 2025-09-07T07:36:00.2591442Z * [new tag] ciflow/xpu/161246 -> ciflow/xpu/161246 2025-09-07T07:36:00.2591825Z * [new tag] ciflow/xpu/161397 -> ciflow/xpu/161397 2025-09-07T07:36:00.2592216Z * [new tag] ciflow/xpu/161485 -> ciflow/xpu/161485 2025-09-07T07:36:00.2592585Z * [new tag] ciflow/xpu/161988 -> ciflow/xpu/161988 2025-09-07T07:36:00.2592912Z * [new tag] ciflow/xpu/162062 -> ciflow/xpu/162062 2025-09-07T07:36:00.2593554Z * [new tag] cslpull75 -> cslpull75 2025-09-07T07:36:00.2593858Z * [new tag] cslpull76 -> cslpull76 2025-09-07T07:36:00.2594400Z * [new tag] cslpull77 -> cslpull77 2025-09-07T07:36:00.2594705Z * [new tag] cslpull78 -> cslpull78 2025-09-07T07:36:00.2595306Z * [new tag] cslpull79 -> cslpull79 2025-09-07T07:36:00.2595892Z * [new tag] cslpull80 -> cslpull80 2025-09-07T07:36:00.2596400Z * [new tag] cslpull81 -> cslpull81 2025-09-07T07:36:00.2596761Z * [new tag] cslpull82 -> cslpull82 2025-09-07T07:36:00.2597306Z * [new tag] cslpull83 -> cslpull83 2025-09-07T07:36:00.2597610Z * [new tag] cslpull84 -> cslpull84 2025-09-07T07:36:00.2598114Z * [new tag] cslpull85 -> cslpull85 2025-09-07T07:36:00.2598513Z * [new tag] cslpull86 -> cslpull86 2025-09-07T07:36:00.2599029Z * [new tag] cslpull87 -> cslpull87 2025-09-07T07:36:00.2599425Z * [new tag] cslpull88 -> cslpull88 2025-09-07T07:36:00.2599817Z * [new tag] cslpull89 -> cslpull89 2025-09-07T07:36:00.2600195Z * [new tag] cslpull90 -> cslpull90 2025-09-07T07:36:00.2601111Z * [new tag] cslpull91 -> cslpull91 2025-09-07T07:36:00.2601420Z * [new tag] cslpull92 -> cslpull92 2025-09-07T07:36:00.2601832Z * [new tag] flight_5 -> flight_5 2025-09-07T07:36:00.2602356Z * [new tag] flight_5.1 -> flight_5.1 2025-09-07T07:36:00.2602755Z * [new tag] flight_5.2 -> flight_5.2 2025-09-07T07:36:00.2603563Z * [new tag] flight_5.3 -> flight_5.3 2025-09-07T07:36:00.2603926Z * [new tag] forpull1 -> forpull1 2025-09-07T07:36:00.2604540Z * [new tag] malfet/tag-2ef5611 -> malfet/tag-2ef5611 2025-09-07T07:36:00.2604899Z * [new tag] malfet/tag-317b1a0 -> malfet/tag-317b1a0 2025-09-07T07:36:00.2605421Z * [new tag] malfet/tag-ec6f767 -> malfet/tag-ec6f767 2025-09-07T07:36:00.2605829Z * [new tag] nightly-binary -> nightly-binary 2025-09-07T07:36:00.2606146Z * [new tag] sqzhang_flight4_plus -> sqzhang_flight4_plus 2025-09-07T07:36:00.2606757Z * [new tag] sqzhang_flight_3 -> sqzhang_flight_3 2025-09-07T07:36:00.2607528Z * [new tag] trunk/00636e0171e7e733628c408084805442270cf608 -> trunk/00636e0171e7e733628c408084805442270cf608 2025-09-07T07:36:00.2607967Z * [new tag] trunk/019fed39aa6b2dd8c69347378d53423e5efae8d4 -> trunk/019fed39aa6b2dd8c69347378d53423e5efae8d4 2025-09-07T07:36:00.2608643Z * [new tag] trunk/01ab325cc2e0dc221af4d710974e1b9175066544 -> trunk/01ab325cc2e0dc221af4d710974e1b9175066544 2025-09-07T07:36:00.2609109Z * [new tag] trunk/01edcd4df8bf0c7b4cc2d3ec868bd2059eeea83b -> trunk/01edcd4df8bf0c7b4cc2d3ec868bd2059eeea83b 2025-09-07T07:36:00.2609581Z * [new tag] trunk/040d00af048967dde7938d358d7f5988cbd18388 -> trunk/040d00af048967dde7938d358d7f5988cbd18388 2025-09-07T07:36:00.2610067Z * [new tag] trunk/0447f2d99b4351b2ff129dce6eebb371024f73e5 -> trunk/0447f2d99b4351b2ff129dce6eebb371024f73e5 2025-09-07T07:36:00.2610557Z * [new tag] trunk/047603d35bdc70046216384838d6340feab79bf4 -> trunk/047603d35bdc70046216384838d6340feab79bf4 2025-09-07T07:36:00.2611045Z * [new tag] trunk/06da7c0730b3764f178ec3a90dedf4ffa4202d81 -> trunk/06da7c0730b3764f178ec3a90dedf4ffa4202d81 2025-09-07T07:36:00.2611717Z * [new tag] trunk/081cab045472ce045634548cc6c14a4870641e23 -> trunk/081cab045472ce045634548cc6c14a4870641e23 2025-09-07T07:36:00.2612060Z * [new tag] trunk/09587daf8c9f21f5340f73921ce5f23d1a4a4572 -> trunk/09587daf8c9f21f5340f73921ce5f23d1a4a4572 2025-09-07T07:36:00.2612498Z * [new tag] trunk/09be1890d72cc34fc946965dc4a27736bf0ca8c6 -> trunk/09be1890d72cc34fc946965dc4a27736bf0ca8c6 2025-09-07T07:36:00.2612938Z * [new tag] trunk/09d2f1b6315d6d416fbf452793d65795863ebc66 -> trunk/09d2f1b6315d6d416fbf452793d65795863ebc66 2025-09-07T07:36:00.2613345Z * [new tag] trunk/0af70e2353e1dcda83175fd4834ecb7b63e009e0 -> trunk/0af70e2353e1dcda83175fd4834ecb7b63e009e0 2025-09-07T07:36:00.2614370Z * [new tag] trunk/0c0e056a9e20c17271a6144dd32c0c7e3ba26736 -> trunk/0c0e056a9e20c17271a6144dd32c0c7e3ba26736 2025-09-07T07:36:00.2614717Z * [new tag] trunk/0cd6c56bdfa9178ff61be82ce3b178926ddb64a9 -> trunk/0cd6c56bdfa9178ff61be82ce3b178926ddb64a9 2025-09-07T07:36:00.2615175Z * [new tag] trunk/0d421ace32c1605ee8e452ee1eeb03bd243dd96c -> trunk/0d421ace32c1605ee8e452ee1eeb03bd243dd96c 2025-09-07T07:36:00.2616110Z * [new tag] trunk/0d71a9dd5b4b6d1dde58d91c9b71d96bc6a6a171 -> trunk/0d71a9dd5b4b6d1dde58d91c9b71d96bc6a6a171 2025-09-07T07:36:00.2616432Z * [new tag] trunk/0d84ff3b78f55492d3d4708458c92d776274939e -> trunk/0d84ff3b78f55492d3d4708458c92d776274939e 2025-09-07T07:36:00.2616933Z * [new tag] trunk/0f45aaf4414048b17d720d0915ce221a8de8ec63 -> trunk/0f45aaf4414048b17d720d0915ce221a8de8ec63 2025-09-07T07:36:00.2617385Z * [new tag] trunk/0ff8eabf1387de5acd6712a03bda61f1a3dfa27f -> trunk/0ff8eabf1387de5acd6712a03bda61f1a3dfa27f 2025-09-07T07:36:00.2617801Z * [new tag] trunk/104f2680e03d13a4765ca69f905d8f16fc0c822f -> trunk/104f2680e03d13a4765ca69f905d8f16fc0c822f 2025-09-07T07:36:00.2618299Z * [new tag] trunk/12814701555d3e41dfcdf8f9273af5821e322df0 -> trunk/12814701555d3e41dfcdf8f9273af5821e322df0 2025-09-07T07:36:00.2618931Z * [new tag] trunk/13b65196db422bdb394cb482e208c61ed448898c -> trunk/13b65196db422bdb394cb482e208c61ed448898c 2025-09-07T07:36:00.2619257Z * [new tag] trunk/13d66e2a66eceed14b8a8f5a971087df4f688a46 -> trunk/13d66e2a66eceed14b8a8f5a971087df4f688a46 2025-09-07T07:36:00.2619695Z * [new tag] trunk/145a3a7bda15e3963a33eb1b54bba5d4a270b225 -> trunk/145a3a7bda15e3963a33eb1b54bba5d4a270b225 2025-09-07T07:36:00.2620172Z * [new tag] trunk/146371483318e17929daefd37c8e459d9d6d47bb -> trunk/146371483318e17929daefd37c8e459d9d6d47bb 2025-09-07T07:36:00.2620618Z * [new tag] trunk/15c77a8cfd341e74fd124b077492ef2bfa51b339 -> trunk/15c77a8cfd341e74fd124b077492ef2bfa51b339 2025-09-07T07:36:00.2621070Z * [new tag] trunk/17fa8eec4a1e32939ab4d364ee6e75487a79b654 -> trunk/17fa8eec4a1e32939ab4d364ee6e75487a79b654 2025-09-07T07:36:00.2622038Z * [new tag] trunk/190c391a28845a14df26abb228d26aa813efb20c -> trunk/190c391a28845a14df26abb228d26aa813efb20c 2025-09-07T07:36:00.2622759Z * [new tag] trunk/1a588ace4667bde1331fbd8ed957157dca5cee68 -> trunk/1a588ace4667bde1331fbd8ed957157dca5cee68 2025-09-07T07:36:00.2623105Z * [new tag] trunk/1aa7476885e8f6e7b0ec3a5b6383aad9d3f343e7 -> trunk/1aa7476885e8f6e7b0ec3a5b6383aad9d3f343e7 2025-09-07T07:36:00.2623516Z * [new tag] trunk/1aeb421c342c9e9607842f4c87cb46e8e816ee53 -> trunk/1aeb421c342c9e9607842f4c87cb46e8e816ee53 2025-09-07T07:36:00.2623946Z * [new tag] trunk/1c1b28d5b6a942fafe23b2f09302d93c25226d4a -> trunk/1c1b28d5b6a942fafe23b2f09302d93c25226d4a 2025-09-07T07:36:00.2624398Z * [new tag] trunk/1ebd70d0c0d562d3be9abdee2a21906584af7d99 -> trunk/1ebd70d0c0d562d3be9abdee2a21906584af7d99 2025-09-07T07:36:00.2624812Z * [new tag] trunk/1ec2c15914da4ef7bd926ed9aebc8671c75fe965 -> trunk/1ec2c15914da4ef7bd926ed9aebc8671c75fe965 2025-09-07T07:36:00.2625228Z * [new tag] trunk/1f51056bd64e73d1aa81321bc3c098575b1bc78a -> trunk/1f51056bd64e73d1aa81321bc3c098575b1bc78a 2025-09-07T07:36:00.2625737Z * [new tag] trunk/1f820de639c75a1562d3fb03f160439f853ae07b -> trunk/1f820de639c75a1562d3fb03f160439f853ae07b 2025-09-07T07:36:00.2626175Z * [new tag] trunk/204697f0e695d82894c5010fbec664c4391f90cc -> trunk/204697f0e695d82894c5010fbec664c4391f90cc 2025-09-07T07:36:00.2626621Z * [new tag] trunk/20629b1619fe636227d01fc85ba221daa7185a05 -> trunk/20629b1619fe636227d01fc85ba221daa7185a05 2025-09-07T07:36:00.2627047Z * [new tag] trunk/20b47acef845e9c4f71da9429a396d293f50ebe7 -> trunk/20b47acef845e9c4f71da9429a396d293f50ebe7 2025-09-07T07:36:00.2627507Z * [new tag] trunk/20bfb2539d7c5250379648eda35f80b8a7d642dd -> trunk/20bfb2539d7c5250379648eda35f80b8a7d642dd 2025-09-07T07:36:00.2627988Z * [new tag] trunk/21fae99c180d17def562797ea0fb154d8fdf88e3 -> trunk/21fae99c180d17def562797ea0fb154d8fdf88e3 2025-09-07T07:36:00.2628461Z * [new tag] trunk/248355faf53f9f7ba2fd0a367d59600c6d991e7f -> trunk/248355faf53f9f7ba2fd0a367d59600c6d991e7f 2025-09-07T07:36:00.2628837Z * [new tag] trunk/25f4aaed9ec26f39c13862323ff8582006473d23 -> trunk/25f4aaed9ec26f39c13862323ff8582006473d23 2025-09-07T07:36:00.2629285Z * [new tag] trunk/261a84a1764412f8e659c956e3f81997ec3de9d5 -> trunk/261a84a1764412f8e659c956e3f81997ec3de9d5 2025-09-07T07:36:00.2629904Z * [new tag] trunk/28f4ab0737937858730f29f5c4e601e109cf9d5f -> trunk/28f4ab0737937858730f29f5c4e601e109cf9d5f 2025-09-07T07:36:00.2630304Z * [new tag] trunk/291cd11f2d5df6f48d348cce0e4e762f274f4dc4 -> trunk/291cd11f2d5df6f48d348cce0e4e762f274f4dc4 2025-09-07T07:36:00.2630768Z * [new tag] trunk/29280864d941e6108ab57f7298f520c0cf9696e9 -> trunk/29280864d941e6108ab57f7298f520c0cf9696e9 2025-09-07T07:36:00.2631233Z * [new tag] trunk/2a45837e98c63cae9d1a2e2133a727b829e549d5 -> trunk/2a45837e98c63cae9d1a2e2133a727b829e549d5 2025-09-07T07:36:00.2631703Z * [new tag] trunk/2a5c0785e2f975697fd7bdf1411de6e03dcaa1ef -> trunk/2a5c0785e2f975697fd7bdf1411de6e03dcaa1ef 2025-09-07T07:36:00.2632229Z * [new tag] trunk/2b8a83901c58a0858ea9e4ce00055f48e6ed164c -> trunk/2b8a83901c58a0858ea9e4ce00055f48e6ed164c 2025-09-07T07:36:00.2632638Z * [new tag] trunk/2ba65472dd54488a86a50326ea990195fc6732d6 -> trunk/2ba65472dd54488a86a50326ea990195fc6732d6 2025-09-07T07:36:00.2633093Z * [new tag] trunk/2c03f0acc53ed13fe8ebfe809129f25996e009a0 -> trunk/2c03f0acc53ed13fe8ebfe809129f25996e009a0 2025-09-07T07:36:00.2633512Z * [new tag] trunk/2dd529df0092799f68ee7afcf52338276906706a -> trunk/2dd529df0092799f68ee7afcf52338276906706a 2025-09-07T07:36:00.2634146Z * [new tag] trunk/2f6b4b1ad3f82bb3bd984f6e65744ea339ffb8b5 -> trunk/2f6b4b1ad3f82bb3bd984f6e65744ea339ffb8b5 2025-09-07T07:36:00.2634466Z * [new tag] trunk/2fa0520a64ed8aa734a56c4d124958f0b5711ca8 -> trunk/2fa0520a64ed8aa734a56c4d124958f0b5711ca8 2025-09-07T07:36:00.2634923Z * [new tag] trunk/302df2ac5dc4222294c09d48804a2dddb8f4bad8 -> trunk/302df2ac5dc4222294c09d48804a2dddb8f4bad8 2025-09-07T07:36:00.2635259Z * [new tag] trunk/33028597bfa2e0178e28c8cce33cb9b3800cac43 -> trunk/33028597bfa2e0178e28c8cce33cb9b3800cac43 2025-09-07T07:36:00.2635939Z * [new tag] trunk/34aa78274d6770086025a967fa63a86830e08176 -> trunk/34aa78274d6770086025a967fa63a86830e08176 2025-09-07T07:36:00.2636323Z * [new tag] trunk/3559c354ce6a14d11fe29fb12fa2747a2f2af449 -> trunk/3559c354ce6a14d11fe29fb12fa2747a2f2af449 2025-09-07T07:36:00.2636695Z * [new tag] trunk/36d207fcaaede0d1e58a5168084c307b32b6fd8b -> trunk/36d207fcaaede0d1e58a5168084c307b32b6fd8b 2025-09-07T07:36:00.2637049Z * [new tag] trunk/377033757ae5ca524ea842f1b0a5f446ed3d8fe0 -> trunk/377033757ae5ca524ea842f1b0a5f446ed3d8fe0 2025-09-07T07:36:00.2637477Z * [new tag] trunk/3771380f83fcac154a7c89ad679311d8c4818287 -> trunk/3771380f83fcac154a7c89ad679311d8c4818287 2025-09-07T07:36:00.2637939Z * [new tag] trunk/3a207816cc569f78863d86c01f2a3d265350e39f -> trunk/3a207816cc569f78863d86c01f2a3d265350e39f 2025-09-07T07:36:00.2638337Z * [new tag] trunk/3a20a20e7065ec927fdd216d4da3b04f879b3c67 -> trunk/3a20a20e7065ec927fdd216d4da3b04f879b3c67 2025-09-07T07:36:00.2638852Z * [new tag] trunk/3bbc2e3e4f025523eaa5dbff220b3e96bca608d0 -> trunk/3bbc2e3e4f025523eaa5dbff220b3e96bca608d0 2025-09-07T07:36:00.2639262Z * [new tag] trunk/3c0ff1b569c45cfa6935ad8031a9d4cf1551aa3f -> trunk/3c0ff1b569c45cfa6935ad8031a9d4cf1551aa3f 2025-09-07T07:36:00.2639720Z * [new tag] trunk/3c45af079afc92a03b03ddf4f9198902ffcf30cf -> trunk/3c45af079afc92a03b03ddf4f9198902ffcf30cf 2025-09-07T07:36:00.2640319Z * [new tag] trunk/3dde5d7f9bf80dd6623a712bc429e9e4302464b5 -> trunk/3dde5d7f9bf80dd6623a712bc429e9e4302464b5 2025-09-07T07:36:00.2640587Z * [new tag] trunk/403a3a393cda7e60f503f3b04b8805a845dcf45d -> trunk/403a3a393cda7e60f503f3b04b8805a845dcf45d 2025-09-07T07:36:00.2641035Z * [new tag] trunk/420c52ecf36f86d32da0853bfbe074b682b070aa -> trunk/420c52ecf36f86d32da0853bfbe074b682b070aa 2025-09-07T07:36:00.2641504Z * [new tag] trunk/43b7c86a2c0f91320f5c5f4827b111edff06fdb6 -> trunk/43b7c86a2c0f91320f5c5f4827b111edff06fdb6 2025-09-07T07:36:00.2641948Z * [new tag] trunk/451ed931562ec8b46d1f7e6c266a68132a119336 -> trunk/451ed931562ec8b46d1f7e6c266a68132a119336 2025-09-07T07:36:00.2642358Z * [new tag] trunk/480c7391126656154318fabf1d57ebc01e196e63 -> trunk/480c7391126656154318fabf1d57ebc01e196e63 2025-09-07T07:36:00.2642824Z * [new tag] trunk/48bedd753da22634aa94fbafeb731e82025404f3 -> trunk/48bedd753da22634aa94fbafeb731e82025404f3 2025-09-07T07:36:00.2643163Z * [new tag] trunk/494878a11b79071ada0b98f34042d47155be6d1c -> trunk/494878a11b79071ada0b98f34042d47155be6d1c 2025-09-07T07:36:00.2643641Z * [new tag] trunk/4ae57d448c0a7d37e4cfd5c27d977fad2cef4051 -> trunk/4ae57d448c0a7d37e4cfd5c27d977fad2cef4051 2025-09-07T07:36:00.2644091Z * [new tag] trunk/4cdaf8265d86f984254b62052da8c26ef61ef1cf -> trunk/4cdaf8265d86f984254b62052da8c26ef61ef1cf 2025-09-07T07:36:00.2644424Z * [new tag] trunk/4d4abec80f03cd8fdefe1d9cb3a60d3690cd777e -> trunk/4d4abec80f03cd8fdefe1d9cb3a60d3690cd777e 2025-09-07T07:36:00.2644942Z * [new tag] trunk/4e42aa8ffc44b8340eb0eeaf80a2cafc4763a186 -> trunk/4e42aa8ffc44b8340eb0eeaf80a2cafc4763a186 2025-09-07T07:36:00.2645783Z * [new tag] trunk/4f72d932feee0749397fec876dcd43994f50b215 -> trunk/4f72d932feee0749397fec876dcd43994f50b215 2025-09-07T07:36:00.2646237Z * [new tag] trunk/50fc22dedf3c4a27be61fa05551c4f320281b42d -> trunk/50fc22dedf3c4a27be61fa05551c4f320281b42d 2025-09-07T07:36:00.2646735Z * [new tag] trunk/5211f1f908907ffc064b56e43cf8659f7fc22aa9 -> trunk/5211f1f908907ffc064b56e43cf8659f7fc22aa9 2025-09-07T07:36:00.2647205Z * [new tag] trunk/524b78d4f67045b83bb69edc56ab16efe282971c -> trunk/524b78d4f67045b83bb69edc56ab16efe282971c 2025-09-07T07:36:00.2647651Z * [new tag] trunk/54e275e0d81fe1e1ccfa4fb5f2a5a9aaca00ca15 -> trunk/54e275e0d81fe1e1ccfa4fb5f2a5a9aaca00ca15 2025-09-07T07:36:00.2647987Z * [new tag] trunk/5561e45758d59c94605873d5db48ed459c004c3b -> trunk/5561e45758d59c94605873d5db48ed459c004c3b 2025-09-07T07:36:00.2648726Z * [new tag] trunk/57278d45f046d4f89f45d373b1af4dd56934ff24 -> trunk/57278d45f046d4f89f45d373b1af4dd56934ff24 2025-09-07T07:36:00.2649189Z * [new tag] trunk/5927a70934ccf7b70182d364c23245a7dd685503 -> trunk/5927a70934ccf7b70182d364c23245a7dd685503 2025-09-07T07:36:00.2649632Z * [new tag] trunk/5985e28912aeb40b103ebfcf2fd0665eb4a50599 -> trunk/5985e28912aeb40b103ebfcf2fd0665eb4a50599 2025-09-07T07:36:00.2650116Z * [new tag] trunk/5a2da090ed6db88bb657c4e51ec0b310cd08bff6 -> trunk/5a2da090ed6db88bb657c4e51ec0b310cd08bff6 2025-09-07T07:36:00.2650600Z * [new tag] trunk/5c473e9f5ee0ef0fc38e6cf34a95b547f8cdc8d5 -> trunk/5c473e9f5ee0ef0fc38e6cf34a95b547f8cdc8d5 2025-09-07T07:36:00.2651018Z * [new tag] trunk/5c67426d6847667a7c55a2dd01f470fa37238c18 -> trunk/5c67426d6847667a7c55a2dd01f470fa37238c18 2025-09-07T07:36:00.2651459Z * [new tag] trunk/5da573c42c332bc68d4b7946c69f690a876d951a -> trunk/5da573c42c332bc68d4b7946c69f690a876d951a 2025-09-07T07:36:00.2651920Z * [new tag] trunk/5e5870e858f60ff4bf87d03f3592097e934a9580 -> trunk/5e5870e858f60ff4bf87d03f3592097e934a9580 2025-09-07T07:36:00.2652333Z * [new tag] trunk/5f3cbc9442aa55b5afb29f4ac8ca9be569003e84 -> trunk/5f3cbc9442aa55b5afb29f4ac8ca9be569003e84 2025-09-07T07:36:00.2652815Z * [new tag] trunk/600c25e9a17fe56e3dee872be8854db08916ba0c -> trunk/600c25e9a17fe56e3dee872be8854db08916ba0c 2025-09-07T07:36:00.2653223Z * [new tag] trunk/601ae8e4831fc8123fffcfb8fd2e6b6381b42e14 -> trunk/601ae8e4831fc8123fffcfb8fd2e6b6381b42e14 2025-09-07T07:36:00.2653692Z * [new tag] trunk/6087ef41e54c2494b117ffd923faf20f515a6806 -> trunk/6087ef41e54c2494b117ffd923faf20f515a6806 2025-09-07T07:36:00.2654148Z * [new tag] trunk/626cb7df8161dd4ecb4fe43b60f37ce9076f56b1 -> trunk/626cb7df8161dd4ecb4fe43b60f37ce9076f56b1 2025-09-07T07:36:00.2654591Z * [new tag] trunk/62c3f9a97fd3dea7132a93066d32d893ffe101e6 -> trunk/62c3f9a97fd3dea7132a93066d32d893ffe101e6 2025-09-07T07:36:00.2655033Z * [new tag] trunk/63a9c23fe99eacfd09610c36dfe8f01b053c1a35 -> trunk/63a9c23fe99eacfd09610c36dfe8f01b053c1a35 2025-09-07T07:36:00.2655485Z * [new tag] trunk/65985937d97505f648b6ed852c3129f2dd08b251 -> trunk/65985937d97505f648b6ed852c3129f2dd08b251 2025-09-07T07:36:00.2656426Z * [new tag] trunk/66f3b4a682a6153517dd23369fdc3289b6494b07 -> trunk/66f3b4a682a6153517dd23369fdc3289b6494b07 2025-09-07T07:36:00.2656674Z * [new tag] trunk/6737e2c996990024187ba620d2764f3b6f6add2c -> trunk/6737e2c996990024187ba620d2764f3b6f6add2c 2025-09-07T07:36:00.2657152Z * [new tag] trunk/67c31dcd364f10072a55f4a30ffd1151c686283a -> trunk/67c31dcd364f10072a55f4a30ffd1151c686283a 2025-09-07T07:36:00.2657634Z * [new tag] trunk/68738beff73e9c3512e18b4edea811a897ce42db -> trunk/68738beff73e9c3512e18b4edea811a897ce42db 2025-09-07T07:36:00.2658115Z * [new tag] trunk/69a25f68884a168550695fdb1a7c310c54d29536 -> trunk/69a25f68884a168550695fdb1a7c310c54d29536 2025-09-07T07:36:00.2658494Z * [new tag] trunk/6b1900c22f1a07b9519346898d4c71d8a2b0f12f -> trunk/6b1900c22f1a07b9519346898d4c71d8a2b0f12f 2025-09-07T07:36:00.2658927Z * [new tag] trunk/6b8b3ac4403f771bd4a8f9a45d93347304148774 -> trunk/6b8b3ac4403f771bd4a8f9a45d93347304148774 2025-09-07T07:36:00.2659403Z * [new tag] trunk/6f7608d603834d6068b2e7a5d59bec3973b6bb1b -> trunk/6f7608d603834d6068b2e7a5d59bec3973b6bb1b 2025-09-07T07:36:00.2659879Z * [new tag] trunk/70d36e047dfb3488fd6335016711a784d810ebda -> trunk/70d36e047dfb3488fd6335016711a784d810ebda 2025-09-07T07:36:00.2660299Z * [new tag] trunk/71992dd805ff9d6763f77214dfe8b0465e88c87b -> trunk/71992dd805ff9d6763f77214dfe8b0465e88c87b 2025-09-07T07:36:00.2660742Z * [new tag] trunk/734ce8eba9c69381f187359bf0fef1d71d84cd20 -> trunk/734ce8eba9c69381f187359bf0fef1d71d84cd20 2025-09-07T07:36:00.2661245Z * [new tag] trunk/73eb4511fb863a37944342b7e92aae706de603c8 -> trunk/73eb4511fb863a37944342b7e92aae706de603c8 2025-09-07T07:36:00.2661747Z * [new tag] trunk/75bc23cfc345bd4c05e7f97c416c4b3d2d1fa64b -> trunk/75bc23cfc345bd4c05e7f97c416c4b3d2d1fa64b 2025-09-07T07:36:00.2662186Z * [new tag] trunk/771f369448321a387f2018535bc8b8b6e5f12fab -> trunk/771f369448321a387f2018535bc8b8b6e5f12fab 2025-09-07T07:36:00.2662897Z * [new tag] trunk/789d4942127143f2adcb53612c058ce4c9a2cf20 -> trunk/789d4942127143f2adcb53612c058ce4c9a2cf20 2025-09-07T07:36:00.2663668Z * [new tag] trunk/791eff96c85678c950888f9da24650083ee673fe -> trunk/791eff96c85678c950888f9da24650083ee673fe 2025-09-07T07:36:00.2663895Z * [new tag] trunk/793fc12aff1f69fbbf9f4278182fb52bbe350fc9 -> trunk/793fc12aff1f69fbbf9f4278182fb52bbe350fc9 2025-09-07T07:36:00.2664108Z * [new tag] trunk/79fcd5247a9a129eee526a14df30bfc6a22b3f01 -> trunk/79fcd5247a9a129eee526a14df30bfc6a22b3f01 2025-09-07T07:36:00.2664377Z * [new tag] trunk/7f4ff79210eb06924f223ae3a1941ee0e2635348 -> trunk/7f4ff79210eb06924f223ae3a1941ee0e2635348 2025-09-07T07:36:00.2665405Z * [new tag] trunk/8076a185c85112be62be292eb47409c88a585b1c -> trunk/8076a185c85112be62be292eb47409c88a585b1c 2025-09-07T07:36:00.2665625Z * [new tag] trunk/80dd397f1979371a5583fa3d5c7352029522a78d -> trunk/80dd397f1979371a5583fa3d5c7352029522a78d 2025-09-07T07:36:00.2665830Z * [new tag] trunk/8171d6052ec12628eb67e0040839314056014429 -> trunk/8171d6052ec12628eb67e0040839314056014429 2025-09-07T07:36:00.2666509Z * [new tag] trunk/81aeefa657b7ccc26b275c50a9f33b2f056e8071 -> trunk/81aeefa657b7ccc26b275c50a9f33b2f056e8071 2025-09-07T07:36:00.2666718Z * [new tag] trunk/81b7b16618bda250ce55982894a83dc0805eb64c -> trunk/81b7b16618bda250ce55982894a83dc0805eb64c 2025-09-07T07:36:00.2667223Z * [new tag] trunk/827f0d405448de31f79d1089f7d7fceab2f87895 -> trunk/827f0d405448de31f79d1089f7d7fceab2f87895 2025-09-07T07:36:00.2667555Z * [new tag] trunk/82f63c8f6de63c30132a8ac299b6e8c2fd0d3fe8 -> trunk/82f63c8f6de63c30132a8ac299b6e8c2fd0d3fe8 2025-09-07T07:36:00.2668001Z * [new tag] trunk/850e1382a9c56bfde18af09d3e72352d775e9435 -> trunk/850e1382a9c56bfde18af09d3e72352d775e9435 2025-09-07T07:36:00.2668521Z * [new tag] trunk/8678d831c48e616b717bff50f2d03141d2e9f965 -> trunk/8678d831c48e616b717bff50f2d03141d2e9f965 2025-09-07T07:36:00.2669022Z * [new tag] trunk/869cbcc16e489a4f5a14a93d5779b0ea86061c60 -> trunk/869cbcc16e489a4f5a14a93d5779b0ea86061c60 2025-09-07T07:36:00.2669589Z * [new tag] trunk/8703debf669bc2238211bfd039f4ecdd8228b7f7 -> trunk/8703debf669bc2238211bfd039f4ecdd8228b7f7 2025-09-07T07:36:00.2670061Z * [new tag] trunk/874069fbe46e82da5cfa405e6c0deb12e89ff608 -> trunk/874069fbe46e82da5cfa405e6c0deb12e89ff608 2025-09-07T07:36:00.2670564Z * [new tag] trunk/8875d6e394da2fffd04f31b28bf258c94d4776a3 -> trunk/8875d6e394da2fffd04f31b28bf258c94d4776a3 2025-09-07T07:36:00.2671063Z * [new tag] trunk/88d94d17e8c5155451393afa6eb3bab48ab61c16 -> trunk/88d94d17e8c5155451393afa6eb3bab48ab61c16 2025-09-07T07:36:00.2671595Z * [new tag] trunk/890626632def7e0ef95a2d01e87a0e4627824a9f -> trunk/890626632def7e0ef95a2d01e87a0e4627824a9f 2025-09-07T07:36:00.2672139Z * [new tag] trunk/8975cda2520b7b1b5bc3b4d8213edf261fa82570 -> trunk/8975cda2520b7b1b5bc3b4d8213edf261fa82570 2025-09-07T07:36:00.2672604Z * [new tag] trunk/89d41d3f61d04f14730ec26f008a59bef6624610 -> trunk/89d41d3f61d04f14730ec26f008a59bef6624610 2025-09-07T07:36:00.2673049Z * [new tag] trunk/8bb213b6d599ef1273fe52f9b1f6d476056c3a41 -> trunk/8bb213b6d599ef1273fe52f9b1f6d476056c3a41 2025-09-07T07:36:00.2673574Z * [new tag] trunk/8e23a1227b5fb2e39afaa7d57c075a75b640a5af -> trunk/8e23a1227b5fb2e39afaa7d57c075a75b640a5af 2025-09-07T07:36:00.2674380Z * [new tag] trunk/8ec551bb354ab2b85fbbba9d461740a20366d248 -> trunk/8ec551bb354ab2b85fbbba9d461740a20366d248 2025-09-07T07:36:00.2674839Z * [new tag] trunk/8fd3c9ce919c8d5c645fd348bba517e948cbc29d -> trunk/8fd3c9ce919c8d5c645fd348bba517e948cbc29d 2025-09-07T07:36:00.2675416Z * [new tag] trunk/90f50f7e68e120d9574e6e3189e37b4280010ad9 -> trunk/90f50f7e68e120d9574e6e3189e37b4280010ad9 2025-09-07T07:36:00.2675897Z * [new tag] trunk/91f0bcf43fc0bc743350d491ac63b77e92054ac9 -> trunk/91f0bcf43fc0bc743350d491ac63b77e92054ac9 2025-09-07T07:36:00.2676791Z * [new tag] trunk/92576a594b8121f6b0b1b5a3ea16d08792fc68ab -> trunk/92576a594b8121f6b0b1b5a3ea16d08792fc68ab 2025-09-07T07:36:00.2677342Z * [new tag] trunk/92a43025e0baa1f2ce345f28d22913b518a1ab9d -> trunk/92a43025e0baa1f2ce345f28d22913b518a1ab9d 2025-09-07T07:36:00.2677730Z * [new tag] trunk/93fb23d6fae7c4e82c4239a1033e522088742634 -> trunk/93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:36:00.2678159Z * [new tag] trunk/9458d1ac3bd70c2af316a8ba95d2c6c9c1199c9c -> trunk/9458d1ac3bd70c2af316a8ba95d2c6c9c1199c9c 2025-09-07T07:36:00.2678728Z * [new tag] trunk/9480cdc0b61488c89a23c2f64f43b2dcedc8728e -> trunk/9480cdc0b61488c89a23c2f64f43b2dcedc8728e 2025-09-07T07:36:00.2679158Z * [new tag] trunk/9491d289b329e4ba4a9f5f5b1be7960671bb7840 -> trunk/9491d289b329e4ba4a9f5f5b1be7960671bb7840 2025-09-07T07:36:00.2679673Z * [new tag] trunk/9499c8761cd2067feb9877414e818f6fd00290f1 -> trunk/9499c8761cd2067feb9877414e818f6fd00290f1 2025-09-07T07:36:00.2680155Z * [new tag] trunk/95ee0bfea99d3d346d6502b91b497d2b35795504 -> trunk/95ee0bfea99d3d346d6502b91b497d2b35795504 2025-09-07T07:36:00.2680601Z * [new tag] trunk/98374612fc2febd686be20761e56bdc2424bc36a -> trunk/98374612fc2febd686be20761e56bdc2424bc36a 2025-09-07T07:36:00.2681218Z * [new tag] trunk/98efc9e93d8fc61eb53cb91378443617cb550500 -> trunk/98efc9e93d8fc61eb53cb91378443617cb550500 2025-09-07T07:36:00.2681707Z * [new tag] trunk/994f2a5dbcbdc915da39bf6f6ce4d1f5e74835c9 -> trunk/994f2a5dbcbdc915da39bf6f6ce4d1f5e74835c9 2025-09-07T07:36:00.2682198Z * [new tag] trunk/99f356fa58c8d726cef022d8710f5491291158f6 -> trunk/99f356fa58c8d726cef022d8710f5491291158f6 2025-09-07T07:36:00.2682675Z * [new tag] trunk/9a1c5c0a078b94d13ac5c1ae0d754d19fb73bf99 -> trunk/9a1c5c0a078b94d13ac5c1ae0d754d19fb73bf99 2025-09-07T07:36:00.2683186Z * [new tag] trunk/9a665ca3c472384e9d722bddba79e5a7680f1abd -> trunk/9a665ca3c472384e9d722bddba79e5a7680f1abd 2025-09-07T07:36:00.2683625Z * [new tag] trunk/9aedb3cd87b52160872173c177f61053d97bed57 -> trunk/9aedb3cd87b52160872173c177f61053d97bed57 2025-09-07T07:36:00.2684111Z * [new tag] trunk/9b81fe281da41f2421506339d26b027a468902f4 -> trunk/9b81fe281da41f2421506339d26b027a468902f4 2025-09-07T07:36:00.2684591Z * [new tag] trunk/9bdcee01f86e2969cff1140cdecfca13cb51816e -> trunk/9bdcee01f86e2969cff1140cdecfca13cb51816e 2025-09-07T07:36:00.2685085Z * [new tag] trunk/9c03d6be87eedc06e524e202e07a7e776551a839 -> trunk/9c03d6be87eedc06e524e202e07a7e776551a839 2025-09-07T07:36:00.2685534Z * [new tag] trunk/9c957723a0fedd9c637e63e023a613019e2cab60 -> trunk/9c957723a0fedd9c637e63e023a613019e2cab60 2025-09-07T07:36:00.2686030Z * [new tag] trunk/9e5247f51d81735e5f1e65e80588985fa93bccc5 -> trunk/9e5247f51d81735e5f1e65e80588985fa93bccc5 2025-09-07T07:36:00.2686532Z * [new tag] trunk/9eadb37cdd699f7e8e8177a5227bfeb16184ef26 -> trunk/9eadb37cdd699f7e8e8177a5227bfeb16184ef26 2025-09-07T07:36:00.2687020Z * [new tag] trunk/a00cdc1e4159db73c9ffb3f25e93e55877709a29 -> trunk/a00cdc1e4159db73c9ffb3f25e93e55877709a29 2025-09-07T07:36:00.2687490Z * [new tag] trunk/a02ee4a816d11380c6f564c1aba64d56af5ba705 -> trunk/a02ee4a816d11380c6f564c1aba64d56af5ba705 2025-09-07T07:36:00.2687922Z * [new tag] trunk/a3c7f77e50f900721817934120d60c2361b3c40d -> trunk/a3c7f77e50f900721817934120d60c2361b3c40d 2025-09-07T07:36:00.2688404Z * [new tag] trunk/a3d72b09ae12126a2b7d4a63a45ac100a882a802 -> trunk/a3d72b09ae12126a2b7d4a63a45ac100a882a802 2025-09-07T07:36:00.2688910Z * [new tag] trunk/a3e5466002791da609fcb069155d8ee347baee92 -> trunk/a3e5466002791da609fcb069155d8ee347baee92 2025-09-07T07:36:00.2689388Z * [new tag] trunk/a714437093ed196eee28f7de454cf4c41badc098 -> trunk/a714437093ed196eee28f7de454cf4c41badc098 2025-09-07T07:36:00.2689844Z * [new tag] trunk/a75e8cd27098f290de0b7439685d05ce02e91356 -> trunk/a75e8cd27098f290de0b7439685d05ce02e91356 2025-09-07T07:36:00.2690263Z * [new tag] trunk/a8d6943d36c1c2a5f90d3573460695bad4b623ae -> trunk/a8d6943d36c1c2a5f90d3573460695bad4b623ae 2025-09-07T07:36:00.2690759Z * [new tag] trunk/a918bbad6ab20649ff82eefb48417ecbe96bcb34 -> trunk/a918bbad6ab20649ff82eefb48417ecbe96bcb34 2025-09-07T07:36:00.2691421Z * [new tag] trunk/a99d8d39bc842d6ebc3e368b178e4884d24b056e -> trunk/a99d8d39bc842d6ebc3e368b178e4884d24b056e 2025-09-07T07:36:00.2691838Z * [new tag] trunk/aac1a50a191b4102d566c9c1ea22f06d6c2e3f02 -> trunk/aac1a50a191b4102d566c9c1ea22f06d6c2e3f02 2025-09-07T07:36:00.2692315Z * [new tag] trunk/aad96a202244c7d0d120c04ba8db593edd8c0f92 -> trunk/aad96a202244c7d0d120c04ba8db593edd8c0f92 2025-09-07T07:36:00.2692802Z * [new tag] trunk/ab643e4dbbaf7b663d4237514cbf01af9b11565c -> trunk/ab643e4dbbaf7b663d4237514cbf01af9b11565c 2025-09-07T07:36:00.2693297Z * [new tag] trunk/abc447174cd2cf8591edbc70a9f836f9a5779f47 -> trunk/abc447174cd2cf8591edbc70a9f836f9a5779f47 2025-09-07T07:36:00.2693774Z * [new tag] trunk/acece97c3a9dceb63194e314da93fdf37cf15a0d -> trunk/acece97c3a9dceb63194e314da93fdf37cf15a0d 2025-09-07T07:36:00.2694353Z * [new tag] trunk/adae7f66aacf3f248c3101b858cf98d5809119fa -> trunk/adae7f66aacf3f248c3101b858cf98d5809119fa 2025-09-07T07:36:00.2694874Z * [new tag] trunk/ae0edc133e61e3b16caf0b2ee0ff3f33ab72af4c -> trunk/ae0edc133e61e3b16caf0b2ee0ff3f33ab72af4c 2025-09-07T07:36:00.2695414Z * [new tag] trunk/aed33a8fcbd60b052d4559d261390c5797129c6d -> trunk/aed33a8fcbd60b052d4559d261390c5797129c6d 2025-09-07T07:36:00.2695913Z * [new tag] trunk/b04e922712080a3652e438d05e8bb74e0cd2d238 -> trunk/b04e922712080a3652e438d05e8bb74e0cd2d238 2025-09-07T07:36:00.2696459Z * [new tag] trunk/b0a3e58dd71c1a039ac0ef51e5bd8f704f632f6f -> trunk/b0a3e58dd71c1a039ac0ef51e5bd8f704f632f6f 2025-09-07T07:36:00.2696935Z * [new tag] trunk/b16d3f4c8c01d461c2f01064e9ca5fa2b33f5cf1 -> trunk/b16d3f4c8c01d461c2f01064e9ca5fa2b33f5cf1 2025-09-07T07:36:00.2697414Z * [new tag] trunk/b18bb6796f210a183e687d9d64984a5a9d13cf09 -> trunk/b18bb6796f210a183e687d9d64984a5a9d13cf09 2025-09-07T07:36:00.2697873Z * [new tag] trunk/b1bb98ddebdd3e41bf7987372409bdce96ae55de -> trunk/b1bb98ddebdd3e41bf7987372409bdce96ae55de 2025-09-07T07:36:00.2698323Z * [new tag] trunk/b2b4add0e754411372060e1d7b4057a66439172b -> trunk/b2b4add0e754411372060e1d7b4057a66439172b 2025-09-07T07:36:00.2698835Z * [new tag] trunk/b2c7b9ad2dc5a7c0b61febd307761bd5bc2f0f05 -> trunk/b2c7b9ad2dc5a7c0b61febd307761bd5bc2f0f05 2025-09-07T07:36:00.2699314Z * [new tag] trunk/b40d9432be44a6b5974ee62e7d19c3c61c5ece37 -> trunk/b40d9432be44a6b5974ee62e7d19c3c61c5ece37 2025-09-07T07:36:00.2699786Z * [new tag] trunk/b4ad38279b178b7bd14355123c1101e2e853e77b -> trunk/b4ad38279b178b7bd14355123c1101e2e853e77b 2025-09-07T07:36:00.2700251Z * [new tag] trunk/b67c41039835bd9b20b83cd6233e86baaa5f5dde -> trunk/b67c41039835bd9b20b83cd6233e86baaa5f5dde 2025-09-07T07:36:00.2700870Z * [new tag] trunk/b6d0a9ea9056ede4f7024dbf3bd6c43be3aff49c -> trunk/b6d0a9ea9056ede4f7024dbf3bd6c43be3aff49c 2025-09-07T07:36:00.2701308Z * [new tag] trunk/b7dad7dd49448c88d0751fa2e29c70afe985f734 -> trunk/b7dad7dd49448c88d0751fa2e29c70afe985f734 2025-09-07T07:36:00.2701792Z * [new tag] trunk/b7e207ca9f046ddd716076965a0cce403ba99052 -> trunk/b7e207ca9f046ddd716076965a0cce403ba99052 2025-09-07T07:36:00.2702275Z * [new tag] trunk/b919560c4a7010e2d89facee25586269a994746e -> trunk/b919560c4a7010e2d89facee25586269a994746e 2025-09-07T07:36:00.2702841Z * [new tag] trunk/b9ba612f7a968f7b27e121ca8f4d0a4d954f5354 -> trunk/b9ba612f7a968f7b27e121ca8f4d0a4d954f5354 2025-09-07T07:36:00.2703394Z * [new tag] trunk/ba7f546ccccb5e0b36d9070dc25f26a9647f89f8 -> trunk/ba7f546ccccb5e0b36d9070dc25f26a9647f89f8 2025-09-07T07:36:00.2703852Z * [new tag] trunk/bb950284c7e72905994bc25dd436c10e48088d85 -> trunk/bb950284c7e72905994bc25dd436c10e48088d85 2025-09-07T07:36:00.2704366Z * [new tag] trunk/bbedc71fd3267c639c38b4ec25eaa22f973d9c4d -> trunk/bbedc71fd3267c639c38b4ec25eaa22f973d9c4d 2025-09-07T07:36:00.2704798Z * [new tag] trunk/bc4db2c27fce6ff1648bdc5af31ec225d2a31f37 -> trunk/bc4db2c27fce6ff1648bdc5af31ec225d2a31f37 2025-09-07T07:36:00.2705239Z * [new tag] trunk/bc505977fb66677a09c31155c987330fbb18a865 -> trunk/bc505977fb66677a09c31155c987330fbb18a865 2025-09-07T07:36:00.2705848Z * [new tag] trunk/bd39e47feea7326afb5bbb67fcb1e69279239527 -> trunk/bd39e47feea7326afb5bbb67fcb1e69279239527 2025-09-07T07:36:00.2706384Z * [new tag] trunk/be5b03dde96638f25ffd732a4fed7e41b4cf40e1 -> trunk/be5b03dde96638f25ffd732a4fed7e41b4cf40e1 2025-09-07T07:36:00.2706880Z * [new tag] trunk/bffc7dd1f374d8408911cd22c6b3d6df39ded9b3 -> trunk/bffc7dd1f374d8408911cd22c6b3d6df39ded9b3 2025-09-07T07:36:00.2707393Z * [new tag] trunk/c024b1f5a18d5c5aee5cc2acdd4c52b24b93ffcf -> trunk/c024b1f5a18d5c5aee5cc2acdd4c52b24b93ffcf 2025-09-07T07:36:00.2707854Z * [new tag] trunk/c0983e6cc0acf71689e1851d12609e00b3f59371 -> trunk/c0983e6cc0acf71689e1851d12609e00b3f59371 2025-09-07T07:36:00.2708642Z * [new tag] trunk/c10195e723eeeedd099ed8b73eda7184ca618fad -> trunk/c10195e723eeeedd099ed8b73eda7184ca618fad 2025-09-07T07:36:00.2709105Z * [new tag] trunk/c157cf6488ade6a7ee2ce2d25b059e1335630a99 -> trunk/c157cf6488ade6a7ee2ce2d25b059e1335630a99 2025-09-07T07:36:00.2709607Z * [new tag] trunk/c2a30246172fd71d56529907ffd3c27b76b1f3a7 -> trunk/c2a30246172fd71d56529907ffd3c27b76b1f3a7 2025-09-07T07:36:00.2710093Z * [new tag] trunk/c32111149921b48bfef909293f1049e21619ed76 -> trunk/c32111149921b48bfef909293f1049e21619ed76 2025-09-07T07:36:00.2710496Z * [new tag] trunk/c37103234afc832dcad307e9016230810957c9d5 -> trunk/c37103234afc832dcad307e9016230810957c9d5 2025-09-07T07:36:00.2710989Z * [new tag] trunk/c3ceca2995cd35e1376c4b0704669bff1a81e836 -> trunk/c3ceca2995cd35e1376c4b0704669bff1a81e836 2025-09-07T07:36:00.2711567Z * [new tag] trunk/c3d54dea9febb1236d48d19e5d4876a63f2e20fd -> trunk/c3d54dea9febb1236d48d19e5d4876a63f2e20fd 2025-09-07T07:36:00.2712024Z * [new tag] trunk/c465b3d52c5687fe910d35a5c75341b77f821741 -> trunk/c465b3d52c5687fe910d35a5c75341b77f821741 2025-09-07T07:36:00.2712530Z * [new tag] trunk/c5b8a10be5e89396da916d1069ffcb7135f0372b -> trunk/c5b8a10be5e89396da916d1069ffcb7135f0372b 2025-09-07T07:36:00.2712966Z * [new tag] trunk/c7e41071a08f4045bc11ab60ec366d7357d56e30 -> trunk/c7e41071a08f4045bc11ab60ec366d7357d56e30 2025-09-07T07:36:00.2713483Z * [new tag] trunk/c98ddaca6d2e19ca37aff00c4ff0cda1e9a6ff65 -> trunk/c98ddaca6d2e19ca37aff00c4ff0cda1e9a6ff65 2025-09-07T07:36:00.2713947Z * [new tag] trunk/cb1e31362c7b53acf4ac95b9f8878064c184f03b -> trunk/cb1e31362c7b53acf4ac95b9f8878064c184f03b 2025-09-07T07:36:00.2714436Z * [new tag] trunk/cbfb005f7cce79974795b148e265f594f59477c8 -> trunk/cbfb005f7cce79974795b148e265f594f59477c8 2025-09-07T07:36:00.2715019Z * [new tag] trunk/cc5bdd12401bda835291d2f3cb297132ebdbf358 -> trunk/cc5bdd12401bda835291d2f3cb297132ebdbf358 2025-09-07T07:36:00.2715637Z * [new tag] trunk/cd529b686d54bbaa443f5b310140de48422d96c7 -> trunk/cd529b686d54bbaa443f5b310140de48422d96c7 2025-09-07T07:36:00.2716071Z * [new tag] trunk/cec0ff122815582af5302360aff03676558c5c87 -> trunk/cec0ff122815582af5302360aff03676558c5c87 2025-09-07T07:36:00.2716560Z * [new tag] trunk/d11720efdb563d02cf4f7d324311fb15a755268e -> trunk/d11720efdb563d02cf4f7d324311fb15a755268e 2025-09-07T07:36:00.2717015Z * [new tag] trunk/d1706d9128ae24d9048167e80d3fe5196d19035e -> trunk/d1706d9128ae24d9048167e80d3fe5196d19035e 2025-09-07T07:36:00.2717562Z * [new tag] trunk/d1a15abfdcaef138f2d9e93a9f46be44f30b766d -> trunk/d1a15abfdcaef138f2d9e93a9f46be44f30b766d 2025-09-07T07:36:00.2718191Z * [new tag] trunk/d232a95d4a79404ca05c1f52d37fde7339dcdf49 -> trunk/d232a95d4a79404ca05c1f52d37fde7339dcdf49 2025-09-07T07:36:00.2718670Z * [new tag] trunk/d2d4c8e9b2371c9aacfb771d9402ac7427b9778e -> trunk/d2d4c8e9b2371c9aacfb771d9402ac7427b9778e 2025-09-07T07:36:00.2719132Z * [new tag] trunk/d33840c542b387ab08ba49aa6c45aa9567fd9be7 -> trunk/d33840c542b387ab08ba49aa6c45aa9567fd9be7 2025-09-07T07:36:00.2719591Z * [new tag] trunk/d5643e8f3a648a99636bfa1f2a41d54bd3c0d0f1 -> trunk/d5643e8f3a648a99636bfa1f2a41d54bd3c0d0f1 2025-09-07T07:36:00.2720121Z * [new tag] trunk/d5b38410b5b6cf75c7a7389972777a6497926ee7 -> trunk/d5b38410b5b6cf75c7a7389972777a6497926ee7 2025-09-07T07:36:00.2720646Z * [new tag] trunk/d5e0f4202ba14632e4d14862ace096609e763462 -> trunk/d5e0f4202ba14632e4d14862ace096609e763462 2025-09-07T07:36:00.2721186Z * [new tag] trunk/d636c181f9140a7b59be10b36eae23039fc2bb72 -> trunk/d636c181f9140a7b59be10b36eae23039fc2bb72 2025-09-07T07:36:00.2722165Z * [new tag] trunk/d64718503728001a1e78168fd7f2d4ff23e57285 -> trunk/d64718503728001a1e78168fd7f2d4ff23e57285 2025-09-07T07:36:00.2722655Z * [new tag] trunk/d67c29ad22670320d676b02e394274af34e8e643 -> trunk/d67c29ad22670320d676b02e394274af34e8e643 2025-09-07T07:36:00.2723137Z * [new tag] trunk/d6b74568e2c98ce58ecc145b72ac66d4caf7ce95 -> trunk/d6b74568e2c98ce58ecc145b72ac66d4caf7ce95 2025-09-07T07:36:00.2723652Z * [new tag] trunk/d711f27845abd45007ccab6076649ebd896c2661 -> trunk/d711f27845abd45007ccab6076649ebd896c2661 2025-09-07T07:36:00.2724137Z * [new tag] trunk/d9d6dde0f42d4bcc8c97671ac50d5096c7e500ab -> trunk/d9d6dde0f42d4bcc8c97671ac50d5096c7e500ab 2025-09-07T07:36:00.2724642Z * [new tag] trunk/da4db4b33d1fdd046650cf19fdbac581a19bf2f9 -> trunk/da4db4b33d1fdd046650cf19fdbac581a19bf2f9 2025-09-07T07:36:00.2725016Z * [new tag] trunk/dac8a4b91c01c3bbc96f54e621b1ea4ffdbd29d1 -> trunk/dac8a4b91c01c3bbc96f54e621b1ea4ffdbd29d1 2025-09-07T07:36:00.2725610Z * [new tag] trunk/dbec08729fb9848bebed6048c63831b87170d061 -> trunk/dbec08729fb9848bebed6048c63831b87170d061 2025-09-07T07:36:00.2726008Z * [new tag] trunk/dcf385395d838f38c8dca25913578230dd43099a -> trunk/dcf385395d838f38c8dca25913578230dd43099a 2025-09-07T07:36:00.2726504Z * [new tag] trunk/dd2519abe83ec3c40d4797492434e41fe3b47e17 -> trunk/dd2519abe83ec3c40d4797492434e41fe3b47e17 2025-09-07T07:36:00.2727007Z * [new tag] trunk/dec72ea4b006dd0fbcaaaa106ad273d73807ab9d -> trunk/dec72ea4b006dd0fbcaaaa106ad273d73807ab9d 2025-09-07T07:36:00.2727493Z * [new tag] trunk/e0a62b266c021b910ce6dc02a6c9429210487717 -> trunk/e0a62b266c021b910ce6dc02a6c9429210487717 2025-09-07T07:36:00.2728067Z * [new tag] trunk/e19e02c84c9dcc408375e5cae3b0709c18b99228 -> trunk/e19e02c84c9dcc408375e5cae3b0709c18b99228 2025-09-07T07:36:00.2728562Z * [new tag] trunk/e304ea4e69d3a7deeb7e48c7450c214a4c953937 -> trunk/e304ea4e69d3a7deeb7e48c7450c214a4c953937 2025-09-07T07:36:00.2729057Z * [new tag] trunk/e3068cdb446adefb5a875616ba37a60235391439 -> trunk/e3068cdb446adefb5a875616ba37a60235391439 2025-09-07T07:36:00.2729568Z * [new tag] trunk/e381d4b0205d5f126c1de534f867ba776f7c3ee6 -> trunk/e381d4b0205d5f126c1de534f867ba776f7c3ee6 2025-09-07T07:36:00.2730104Z * [new tag] trunk/e4bd0ff4f8981b805df32ea5b3550621965ea4f2 -> trunk/e4bd0ff4f8981b805df32ea5b3550621965ea4f2 2025-09-07T07:36:00.2730498Z * [new tag] trunk/e532c9d4f1cdcbc1ea9628f55b9813e77847bdc7 -> trunk/e532c9d4f1cdcbc1ea9628f55b9813e77847bdc7 2025-09-07T07:36:00.2730964Z * [new tag] trunk/e92cd9415377403b6e90585e764639e2e0b5973b -> trunk/e92cd9415377403b6e90585e764639e2e0b5973b 2025-09-07T07:36:00.2731515Z * [new tag] trunk/e9481b6617b5576b099d8ca5798111592e9ad090 -> trunk/e9481b6617b5576b099d8ca5798111592e9ad090 2025-09-07T07:36:00.2731883Z * [new tag] trunk/ea1883dfd3e42defe37b11202b878bb76defa087 -> trunk/ea1883dfd3e42defe37b11202b878bb76defa087 2025-09-07T07:36:00.2732392Z * [new tag] trunk/eac3d6f04cfbbebe3d470dacd216da7d4b1f95a8 -> trunk/eac3d6f04cfbbebe3d470dacd216da7d4b1f95a8 2025-09-07T07:36:00.2732824Z * [new tag] trunk/eb18d32bda75189494d955aa001ade15f10333de -> trunk/eb18d32bda75189494d955aa001ade15f10333de 2025-09-07T07:36:00.2733208Z * [new tag] trunk/ef3be6726f7ff4b77c22db10cec5b686f9107ea9 -> trunk/ef3be6726f7ff4b77c22db10cec5b686f9107ea9 2025-09-07T07:36:00.2733705Z * [new tag] trunk/ef8aabd42422725026cb4dbf48aafa9efa226a04 -> trunk/ef8aabd42422725026cb4dbf48aafa9efa226a04 2025-09-07T07:36:00.2734301Z * [new tag] trunk/f00445b43eee57e20bb9316fa796ca23bf73373b -> trunk/f00445b43eee57e20bb9316fa796ca23bf73373b 2025-09-07T07:36:00.2734781Z * [new tag] trunk/f0c391102b754e3b145e8c59231d2df563487e37 -> trunk/f0c391102b754e3b145e8c59231d2df563487e37 2025-09-07T07:36:00.2735529Z * [new tag] trunk/f27985b7e796fb66a1b476284ba42d8cb360a751 -> trunk/f27985b7e796fb66a1b476284ba42d8cb360a751 2025-09-07T07:36:00.2735978Z * [new tag] trunk/f36f285953700f971552083a5da9d0ceacb63bbd -> trunk/f36f285953700f971552083a5da9d0ceacb63bbd 2025-09-07T07:36:00.2736487Z * [new tag] trunk/f3cebec39ebc110e1c8b06e741896585f7892dbb -> trunk/f3cebec39ebc110e1c8b06e741896585f7892dbb 2025-09-07T07:36:00.2736884Z * [new tag] trunk/f4c33cd44acac92c0b451a04da20ebe9370e5b0c -> trunk/f4c33cd44acac92c0b451a04da20ebe9370e5b0c 2025-09-07T07:36:00.2737432Z * [new tag] trunk/f612045ce105f008b2b675e2fc870163babeb2e8 -> trunk/f612045ce105f008b2b675e2fc870163babeb2e8 2025-09-07T07:36:00.2737909Z * [new tag] trunk/f8746b878dfc1e9639d42cbde832e9b9e792c86c -> trunk/f8746b878dfc1e9639d42cbde832e9b9e792c86c 2025-09-07T07:36:00.2738432Z * [new tag] trunk/f8ffa9194e26523e5f976d4a824d5cc58922727c -> trunk/f8ffa9194e26523e5f976d4a824d5cc58922727c 2025-09-07T07:36:00.2738886Z * [new tag] trunk/f981a7fa5230b98974291fdde32fe8488bc5d469 -> trunk/f981a7fa5230b98974291fdde32fe8488bc5d469 2025-09-07T07:36:00.2739391Z * [new tag] trunk/fbf3d2027daabbcb44d0af274b139be2a248a4f7 -> trunk/fbf3d2027daabbcb44d0af274b139be2a248a4f7 2025-09-07T07:36:00.2739983Z * [new tag] trunk/fca2601c9d628e1bd2d75c7318cd22c4e8c832aa -> trunk/fca2601c9d628e1bd2d75c7318cd22c4e8c832aa 2025-09-07T07:36:00.2740448Z * [new tag] trunk/fea20775ad96bdca972a1811d7d3372f368614ab -> trunk/fea20775ad96bdca972a1811d7d3372f368614ab 2025-09-07T07:36:00.2741227Z * [new tag] trunk/fefee081642f87419a21dc852f7167d4640443cd -> trunk/fefee081642f87419a21dc852f7167d4640443cd 2025-09-07T07:36:00.2741471Z * [new tag] v0.1.1 -> v0.1.1 2025-09-07T07:36:00.2741919Z * [new tag] v0.1.10 -> v0.1.10 2025-09-07T07:36:00.2742357Z * [new tag] v0.1.11 -> v0.1.11 2025-09-07T07:36:00.2742758Z * [new tag] v0.1.12 -> v0.1.12 2025-09-07T07:36:00.2743160Z * [new tag] v0.1.2 -> v0.1.2 2025-09-07T07:36:00.2743751Z * [new tag] v0.1.3 -> v0.1.3 2025-09-07T07:36:00.2744111Z * [new tag] v0.1.4 -> v0.1.4 2025-09-07T07:36:00.2744547Z * [new tag] v0.1.5 -> v0.1.5 2025-09-07T07:36:00.2744960Z * [new tag] v0.1.6 -> v0.1.6 2025-09-07T07:36:00.2745339Z * [new tag] v0.1.7 -> v0.1.7 2025-09-07T07:36:00.2745731Z * [new tag] v0.1.8 -> v0.1.8 2025-09-07T07:36:00.2746150Z * [new tag] v0.1.9 -> v0.1.9 2025-09-07T07:36:00.2746567Z * [new tag] v0.2.0 -> v0.2.0 2025-09-07T07:36:00.2746993Z * [new tag] v0.3.0 -> v0.3.0 2025-09-07T07:36:00.2747504Z * [new tag] v0.3.1 -> v0.3.1 2025-09-07T07:36:00.2747892Z * [new tag] v0.4.0 -> v0.4.0 2025-09-07T07:36:00.2748335Z * [new tag] v0.4.1 -> v0.4.1 2025-09-07T07:36:00.2748773Z * [new tag] v1.0.0 -> v1.0.0 2025-09-07T07:36:00.2749376Z * [new tag] v1.0.0a0 -> v1.0.0a0 2025-09-07T07:36:00.2749792Z * [new tag] v1.0.1 -> v1.0.1 2025-09-07T07:36:00.2750254Z * [new tag] v1.0rc0 -> v1.0rc0 2025-09-07T07:36:00.2750583Z * [new tag] v1.0rc1 -> v1.0rc1 2025-09-07T07:36:00.2751007Z * [new tag] v1.1.0 -> v1.1.0 2025-09-07T07:36:00.2751490Z * [new tag] v1.1.0a0 -> v1.1.0a0 2025-09-07T07:36:00.2752133Z * [new tag] v1.10.0 -> v1.10.0 2025-09-07T07:36:00.2752604Z * [new tag] v1.10.0-rc1 -> v1.10.0-rc1 2025-09-07T07:36:00.2753196Z * [new tag] v1.10.0-rc2 -> v1.10.0-rc2 2025-09-07T07:36:00.2753430Z * [new tag] v1.10.0-rc3 -> v1.10.0-rc3 2025-09-07T07:36:00.2753918Z * [new tag] v1.10.1 -> v1.10.1 2025-09-07T07:36:00.2754208Z * [new tag] v1.10.1-rc1 -> v1.10.1-rc1 2025-09-07T07:36:00.2754536Z * [new tag] v1.10.2 -> v1.10.2 2025-09-07T07:36:00.2754854Z * [new tag] v1.10.2-rc1 -> v1.10.2-rc1 2025-09-07T07:36:00.2755441Z * [new tag] v1.11.0 -> v1.11.0 2025-09-07T07:36:00.2755864Z * [new tag] v1.11.0-rc1 -> v1.11.0-rc1 2025-09-07T07:36:00.2756469Z * [new tag] v1.11.0-rc2 -> v1.11.0-rc2 2025-09-07T07:36:00.2756884Z * [new tag] v1.11.0-rc3 -> v1.11.0-rc3 2025-09-07T07:36:00.2757363Z * [new tag] v1.11.0-rc4 -> v1.11.0-rc4 2025-09-07T07:36:00.2758013Z * [new tag] v1.11.0-rc5 -> v1.11.0-rc5 2025-09-07T07:36:00.2758221Z * [new tag] v1.11.0-rc6 -> v1.11.0-rc6 2025-09-07T07:36:00.2758541Z * [new tag] v1.11.0-rc7 -> v1.11.0-rc7 2025-09-07T07:36:00.2759061Z * [new tag] v1.12.0 -> v1.12.0 2025-09-07T07:36:00.2759522Z * [new tag] v1.12.0-rc1 -> v1.12.0-rc1 2025-09-07T07:36:00.2759985Z * [new tag] v1.12.0-rc2 -> v1.12.0-rc2 2025-09-07T07:36:00.2760438Z * [new tag] v1.12.0-rc3 -> v1.12.0-rc3 2025-09-07T07:36:00.2760979Z * [new tag] v1.12.0-rc4 -> v1.12.0-rc4 2025-09-07T07:36:00.2761417Z * [new tag] v1.12.0-rc5 -> v1.12.0-rc5 2025-09-07T07:36:00.2762015Z * [new tag] v1.12.0-rc6 -> v1.12.0-rc6 2025-09-07T07:36:00.2762331Z * [new tag] v1.12.0-rc7 -> v1.12.0-rc7 2025-09-07T07:36:00.2762659Z * [new tag] v1.12.0-rc8 -> v1.12.0-rc8 2025-09-07T07:36:00.2762980Z * [new tag] v1.12.1 -> v1.12.1 2025-09-07T07:36:00.2763667Z * [new tag] v1.12.1-rc1 -> v1.12.1-rc1 2025-09-07T07:36:00.2764033Z * [new tag] v1.12.1-rc2 -> v1.12.1-rc2 2025-09-07T07:36:00.2764655Z * [new tag] v1.12.1-rc3 -> v1.12.1-rc3 2025-09-07T07:36:00.2765272Z * [new tag] v1.12.1-rc4 -> v1.12.1-rc4 2025-09-07T07:36:00.2765420Z * [new tag] v1.12.1-rc5 -> v1.12.1-rc5 2025-09-07T07:36:00.2766002Z * [new tag] v1.13.0 -> v1.13.0 2025-09-07T07:36:00.2766366Z * [new tag] v1.13.0-rc1 -> v1.13.0-rc1 2025-09-07T07:36:00.2766826Z * [new tag] v1.13.0-rc2 -> v1.13.0-rc2 2025-09-07T07:36:00.2767310Z * [new tag] v1.13.0-rc3 -> v1.13.0-rc3 2025-09-07T07:36:00.2768287Z * [new tag] v1.13.0-rc4 -> v1.13.0-rc4 2025-09-07T07:36:00.2768507Z * [new tag] v1.13.0-rc5 -> v1.13.0-rc5 2025-09-07T07:36:00.2771495Z * [new tag] v1.13.0-rc6 -> v1.13.0-rc6 2025-09-07T07:36:00.2771866Z * [new tag] v1.13.1 -> v1.13.1 2025-09-07T07:36:00.2772212Z * [new tag] v1.13.1-rc1 -> v1.13.1-rc1 2025-09-07T07:36:00.2772649Z * [new tag] v1.2.0 -> v1.2.0 2025-09-07T07:36:00.2773093Z * [new tag] v1.2.0a0 -> v1.2.0a0 2025-09-07T07:36:00.2773541Z * [new tag] v1.3.0 -> v1.3.0 2025-09-07T07:36:00.2774054Z * [new tag] v1.3.0a0 -> v1.3.0a0 2025-09-07T07:36:00.2774348Z * [new tag] v1.3.1 -> v1.3.1 2025-09-07T07:36:00.2774933Z * [new tag] v1.4.0 -> v1.4.0 2025-09-07T07:36:00.2775285Z * [new tag] v1.4.0a0 -> v1.4.0a0 2025-09-07T07:36:00.2775599Z * [new tag] v1.4.1 -> v1.4.1 2025-09-07T07:36:00.2776076Z * [new tag] v1.5.0 -> v1.5.0 2025-09-07T07:36:00.2776740Z * [new tag] v1.5.0-rc1 -> v1.5.0-rc1 2025-09-07T07:36:00.2777091Z * [new tag] v1.5.0-rc2 -> v1.5.0-rc2 2025-09-07T07:36:00.2777677Z * [new tag] v1.5.0-rc3 -> v1.5.0-rc3 2025-09-07T07:36:00.2778133Z * [new tag] v1.5.0-rc4 -> v1.5.0-rc4 2025-09-07T07:36:00.2778453Z * [new tag] v1.5.0-rc5 -> v1.5.0-rc5 2025-09-07T07:36:00.2779037Z * [new tag] v1.5.1 -> v1.5.1 2025-09-07T07:36:00.2779286Z * [new tag] v1.5.1-rc1 -> v1.5.1-rc1 2025-09-07T07:36:00.2779585Z * [new tag] v1.6.0 -> v1.6.0 2025-09-07T07:36:00.2780074Z * [new tag] v1.6.0-rc1 -> v1.6.0-rc1 2025-09-07T07:36:00.2780662Z * [new tag] v1.6.0-rc2 -> v1.6.0-rc2 2025-09-07T07:36:00.2781081Z * [new tag] v1.6.0-rc3 -> v1.6.0-rc3 2025-09-07T07:36:00.2781511Z * [new tag] v1.6.0-rc4 -> v1.6.0-rc4 2025-09-07T07:36:00.2781973Z * [new tag] v1.6.0-rc5 -> v1.6.0-rc5 2025-09-07T07:36:00.2782429Z * [new tag] v1.6.0-rc6 -> v1.6.0-rc6 2025-09-07T07:36:00.2782746Z * [new tag] v1.6.0-rc7 -> v1.6.0-rc7 2025-09-07T07:36:00.2783208Z * [new tag] v1.7.0 -> v1.7.0 2025-09-07T07:36:00.2783691Z * [new tag] v1.7.0-rc1 -> v1.7.0-rc1 2025-09-07T07:36:00.2784343Z * [new tag] v1.7.0-rc2 -> v1.7.0-rc2 2025-09-07T07:36:00.2784692Z * [new tag] v1.7.0-rc3 -> v1.7.0-rc3 2025-09-07T07:36:00.2785035Z * [new tag] v1.7.0-rc4 -> v1.7.0-rc4 2025-09-07T07:36:00.2785519Z * [new tag] v1.7.1 -> v1.7.1 2025-09-07T07:36:00.2786191Z * [new tag] v1.7.1-rc1 -> v1.7.1-rc1 2025-09-07T07:36:00.2786557Z * [new tag] v1.7.1-rc2 -> v1.7.1-rc2 2025-09-07T07:36:00.2786870Z * [new tag] v1.7.1-rc3 -> v1.7.1-rc3 2025-09-07T07:36:00.2787344Z * [new tag] v1.8.0 -> v1.8.0 2025-09-07T07:36:00.2787661Z * [new tag] v1.8.0-rc1 -> v1.8.0-rc1 2025-09-07T07:36:00.2788182Z * [new tag] v1.8.0-rc2 -> v1.8.0-rc2 2025-09-07T07:36:00.2788578Z * [new tag] v1.8.0-rc3 -> v1.8.0-rc3 2025-09-07T07:36:00.2789154Z * [new tag] v1.8.0-rc4 -> v1.8.0-rc4 2025-09-07T07:36:00.2789387Z * [new tag] v1.8.0-rc5 -> v1.8.0-rc5 2025-09-07T07:36:00.2789702Z * [new tag] v1.8.1 -> v1.8.1 2025-09-07T07:36:00.2790165Z * [new tag] v1.8.1-rc1 -> v1.8.1-rc1 2025-09-07T07:36:00.2790491Z * [new tag] v1.8.1-rc2 -> v1.8.1-rc2 2025-09-07T07:36:00.2790936Z * [new tag] v1.8.1-rc3 -> v1.8.1-rc3 2025-09-07T07:36:00.2791822Z * [new tag] v1.8.2 -> v1.8.2 2025-09-07T07:36:00.2792080Z * [new tag] v1.8.2-rc1 -> v1.8.2-rc1 2025-09-07T07:36:00.2792654Z * [new tag] v1.9.0 -> v1.9.0 2025-09-07T07:36:00.2793069Z * [new tag] v1.9.0-rc1 -> v1.9.0-rc1 2025-09-07T07:36:00.2793697Z * [new tag] v1.9.0-rc2 -> v1.9.0-rc2 2025-09-07T07:36:00.2794098Z * [new tag] v1.9.0-rc3 -> v1.9.0-rc3 2025-09-07T07:36:00.2794409Z * [new tag] v1.9.0-rc4 -> v1.9.0-rc4 2025-09-07T07:36:00.2794893Z * [new tag] v1.9.1 -> v1.9.1 2025-09-07T07:36:00.2795594Z * [new tag] v1.9.1-rc1 -> v1.9.1-rc1 2025-09-07T07:36:00.2795822Z * [new tag] v1.9.1-rc2 -> v1.9.1-rc2 2025-09-07T07:36:00.2796412Z * [new tag] v2.0.0 -> v2.0.0 2025-09-07T07:36:00.2796764Z * [new tag] v2.0.0-rc1 -> v2.0.0-rc1 2025-09-07T07:36:00.2797360Z * [new tag] v2.0.0-rc2 -> v2.0.0-rc2 2025-09-07T07:36:00.2797763Z * [new tag] v2.0.0-rc3 -> v2.0.0-rc3 2025-09-07T07:36:00.2798212Z * [new tag] v2.0.0-rc4 -> v2.0.0-rc4 2025-09-07T07:36:00.2798861Z * [new tag] v2.0.0-rc5 -> v2.0.0-rc5 2025-09-07T07:36:00.2799455Z * [new tag] v2.0.0-rc6 -> v2.0.0-rc6 2025-09-07T07:36:00.2799855Z * [new tag] v2.0.1 -> v2.0.1 2025-09-07T07:36:00.2800344Z * [new tag] v2.0.1-rc1 -> v2.0.1-rc1 2025-09-07T07:36:00.2800650Z * [new tag] v2.0.1-rc2 -> v2.0.1-rc2 2025-09-07T07:36:00.2801082Z * [new tag] v2.0.1-rc3 -> v2.0.1-rc3 2025-09-07T07:36:00.2801393Z * [new tag] v2.0.1-rc4 -> v2.0.1-rc4 2025-09-07T07:36:00.2802323Z * [new tag] v2.1.0 -> v2.1.0 2025-09-07T07:36:00.2802677Z * [new tag] v2.1.0-rc1 -> v2.1.0-rc1 2025-09-07T07:36:00.2803158Z * [new tag] v2.1.0-rc2 -> v2.1.0-rc2 2025-09-07T07:36:00.2803793Z * [new tag] v2.1.0-rc3 -> v2.1.0-rc3 2025-09-07T07:36:00.2804308Z * [new tag] v2.1.0-rc4 -> v2.1.0-rc4 2025-09-07T07:36:00.2804774Z * [new tag] v2.1.0-rc5 -> v2.1.0-rc5 2025-09-07T07:36:00.2805260Z * [new tag] v2.1.0-rc6 -> v2.1.0-rc6 2025-09-07T07:36:00.2805671Z * [new tag] v2.1.1 -> v2.1.1 2025-09-07T07:36:00.2806256Z * [new tag] v2.1.1-rc1 -> v2.1.1-rc1 2025-09-07T07:36:00.2806600Z * [new tag] v2.1.1-rc2 -> v2.1.1-rc2 2025-09-07T07:36:00.2807244Z * [new tag] v2.1.1-rc3 -> v2.1.1-rc3 2025-09-07T07:36:00.2807657Z * [new tag] v2.1.1-rc4 -> v2.1.1-rc4 2025-09-07T07:36:00.2808081Z * [new tag] v2.1.1-rc5 -> v2.1.1-rc5 2025-09-07T07:36:00.2808401Z * [new tag] v2.1.1-rc6 -> v2.1.1-rc6 2025-09-07T07:36:00.2808868Z * [new tag] v2.1.2 -> v2.1.2 2025-09-07T07:36:00.2809480Z * [new tag] v2.1.2-rc1 -> v2.1.2-rc1 2025-09-07T07:36:00.2809871Z * [new tag] v2.1.2-rc2 -> v2.1.2-rc2 2025-09-07T07:36:00.2810202Z * [new tag] v2.1.2-rc3 -> v2.1.2-rc3 2025-09-07T07:36:00.2810801Z * [new tag] v2.2.0 -> v2.2.0 2025-09-07T07:36:00.2811166Z * [new tag] v2.2.0-rc1 -> v2.2.0-rc1 2025-09-07T07:36:00.2811622Z * [new tag] v2.2.0-rc2 -> v2.2.0-rc2 2025-09-07T07:36:00.2812046Z * [new tag] v2.2.0-rc3 -> v2.2.0-rc3 2025-09-07T07:36:00.2812477Z * [new tag] v2.2.0-rc4 -> v2.2.0-rc4 2025-09-07T07:36:00.2812918Z * [new tag] v2.2.0-rc5 -> v2.2.0-rc5 2025-09-07T07:36:00.2813356Z * [new tag] v2.2.0-rc6 -> v2.2.0-rc6 2025-09-07T07:36:00.2813686Z * [new tag] v2.2.0-rc7 -> v2.2.0-rc7 2025-09-07T07:36:00.2814029Z * [new tag] v2.2.0-rc8 -> v2.2.0-rc8 2025-09-07T07:36:00.2814609Z * [new tag] v2.2.1 -> v2.2.1 2025-09-07T07:36:00.2815084Z * [new tag] v2.2.1-rc1 -> v2.2.1-rc1 2025-09-07T07:36:00.2815444Z * [new tag] v2.2.1-rc2 -> v2.2.1-rc2 2025-09-07T07:36:00.2815767Z * [new tag] v2.2.1-rc3 -> v2.2.1-rc3 2025-09-07T07:36:00.2816098Z * [new tag] v2.2.2 -> v2.2.2 2025-09-07T07:36:00.2816711Z * [new tag] v2.2.2-rc1 -> v2.2.2-rc1 2025-09-07T07:36:00.2816972Z * [new tag] v2.2.2-rc2 -> v2.2.2-rc2 2025-09-07T07:36:00.2817313Z * [new tag] v2.2.2-rc3 -> v2.2.2-rc3 2025-09-07T07:36:00.2817951Z * [new tag] v2.3.0 -> v2.3.0 2025-09-07T07:36:00.2818419Z * [new tag] v2.3.0-rc1 -> v2.3.0-rc1 2025-09-07T07:36:00.2819017Z * [new tag] v2.3.0-rc10 -> v2.3.0-rc10 2025-09-07T07:36:00.2819450Z * [new tag] v2.3.0-rc11 -> v2.3.0-rc11 2025-09-07T07:36:00.2819768Z * [new tag] v2.3.0-rc12 -> v2.3.0-rc12 2025-09-07T07:36:00.2820251Z * [new tag] v2.3.0-rc2 -> v2.3.0-rc2 2025-09-07T07:36:00.2820719Z * [new tag] v2.3.0-rc3 -> v2.3.0-rc3 2025-09-07T07:36:00.2821194Z * [new tag] v2.3.0-rc4 -> v2.3.0-rc4 2025-09-07T07:36:00.2821630Z * [new tag] v2.3.0-rc5 -> v2.3.0-rc5 2025-09-07T07:36:00.2821950Z * [new tag] v2.3.0-rc6 -> v2.3.0-rc6 2025-09-07T07:36:00.2822429Z * [new tag] v2.3.0-rc7 -> v2.3.0-rc7 2025-09-07T07:36:00.2823023Z * [new tag] v2.3.0-rc8 -> v2.3.0-rc8 2025-09-07T07:36:00.2823228Z * [new tag] v2.3.0-rc9 -> v2.3.0-rc9 2025-09-07T07:36:00.2823576Z * [new tag] v2.3.1 -> v2.3.1 2025-09-07T07:36:00.2824012Z * [new tag] v2.3.1-rc1 -> v2.3.1-rc1 2025-09-07T07:36:00.2824489Z * [new tag] v2.3.1-rc2 -> v2.3.1-rc2 2025-09-07T07:36:00.2824950Z * [new tag] v2.3.1-rc3 -> v2.3.1-rc3 2025-09-07T07:36:00.2825425Z * [new tag] v2.4.0 -> v2.4.0 2025-09-07T07:36:00.2825947Z * [new tag] v2.4.0-rc1 -> v2.4.0-rc1 2025-09-07T07:36:00.2826377Z * [new tag] v2.4.0-rc2 -> v2.4.0-rc2 2025-09-07T07:36:00.2826819Z * [new tag] v2.4.0-rc3 -> v2.4.0-rc3 2025-09-07T07:36:00.2827310Z * [new tag] v2.4.0-rc4 -> v2.4.0-rc4 2025-09-07T07:36:00.2827931Z * [new tag] v2.4.0-rc5 -> v2.4.0-rc5 2025-09-07T07:36:00.2828771Z * [new tag] v2.4.0-rc6 -> v2.4.0-rc6 2025-09-07T07:36:00.2829188Z * [new tag] v2.4.0-rc7 -> v2.4.0-rc7 2025-09-07T07:36:00.2829632Z * [new tag] v2.4.0-rc8 -> v2.4.0-rc8 2025-09-07T07:36:00.2830229Z * [new tag] v2.4.0-rc9 -> v2.4.0-rc9 2025-09-07T07:36:00.2830486Z * [new tag] v2.4.1 -> v2.4.1 2025-09-07T07:36:00.2831092Z * [new tag] v2.4.1-rc1 -> v2.4.1-rc1 2025-09-07T07:36:00.2831502Z * [new tag] v2.4.1-rc2 -> v2.4.1-rc2 2025-09-07T07:36:00.2832221Z * [new tag] v2.4.1-rc3 -> v2.4.1-rc3 2025-09-07T07:36:00.2832611Z * [new tag] v2.5.0 -> v2.5.0 2025-09-07T07:36:00.2833107Z * [new tag] v2.5.0-rc1 -> v2.5.0-rc1 2025-09-07T07:36:00.2833512Z * [new tag] v2.5.0-rc10 -> v2.5.0-rc10 2025-09-07T07:36:00.2833966Z * [new tag] v2.5.0-rc2 -> v2.5.0-rc2 2025-09-07T07:36:00.2834415Z * [new tag] v2.5.0-rc3 -> v2.5.0-rc3 2025-09-07T07:36:00.2834886Z * [new tag] v2.5.0-rc4 -> v2.5.0-rc4 2025-09-07T07:36:00.2835462Z * [new tag] v2.5.0-rc5 -> v2.5.0-rc5 2025-09-07T07:36:00.2835911Z * [new tag] v2.5.0-rc6 -> v2.5.0-rc6 2025-09-07T07:36:00.2836488Z * [new tag] v2.5.0-rc7 -> v2.5.0-rc7 2025-09-07T07:36:00.2836874Z * [new tag] v2.5.0-rc8 -> v2.5.0-rc8 2025-09-07T07:36:00.2837514Z * [new tag] v2.5.0-rc9 -> v2.5.0-rc9 2025-09-07T07:36:00.2837747Z * [new tag] v2.5.1 -> v2.5.1 2025-09-07T07:36:00.2838091Z * [new tag] v2.5.1-rc1 -> v2.5.1-rc1 2025-09-07T07:36:00.2838405Z * [new tag] v2.6.0 -> v2.6.0 2025-09-07T07:36:00.2839005Z * [new tag] v2.6.0-rc1 -> v2.6.0-rc1 2025-09-07T07:36:00.2839457Z * [new tag] v2.6.0-rc2 -> v2.6.0-rc2 2025-09-07T07:36:00.2840046Z * [new tag] v2.6.0-rc3 -> v2.6.0-rc3 2025-09-07T07:36:00.2840352Z * [new tag] v2.6.0-rc4 -> v2.6.0-rc4 2025-09-07T07:36:00.2841080Z * [new tag] v2.6.0-rc5 -> v2.6.0-rc5 2025-09-07T07:36:00.2841669Z * [new tag] v2.6.0-rc6 -> v2.6.0-rc6 2025-09-07T07:36:00.2842109Z * [new tag] v2.6.0-rc7 -> v2.6.0-rc7 2025-09-07T07:36:00.2842787Z * [new tag] v2.6.0-rc8 -> v2.6.0-rc8 2025-09-07T07:36:00.2843168Z * [new tag] v2.6.0-rc9 -> v2.6.0-rc9 2025-09-07T07:36:00.2844043Z * [new tag] v2.7.0 -> v2.7.0 2025-09-07T07:36:00.2844315Z * [new tag] v2.7.0-rc1 -> v2.7.0-rc1 2025-09-07T07:36:00.2844692Z * [new tag] v2.7.0-rc10 -> v2.7.0-rc10 2025-09-07T07:36:00.2845329Z * [new tag] v2.7.0-rc2 -> v2.7.0-rc2 2025-09-07T07:36:00.2845703Z * [new tag] v2.7.0-rc3 -> v2.7.0-rc3 2025-09-07T07:36:00.2847091Z * [new tag] v2.7.0-rc4 -> v2.7.0-rc4 2025-09-07T07:36:00.2847220Z * [new tag] v2.7.0-rc5 -> v2.7.0-rc5 2025-09-07T07:36:00.2847322Z * [new tag] v2.7.0-rc6 -> v2.7.0-rc6 2025-09-07T07:36:00.2847797Z * [new tag] v2.7.0-rc7 -> v2.7.0-rc7 2025-09-07T07:36:00.2848404Z * [new tag] v2.7.0-rc8 -> v2.7.0-rc8 2025-09-07T07:36:00.2848887Z * [new tag] v2.7.0-rc9 -> v2.7.0-rc9 2025-09-07T07:36:00.2849255Z * [new tag] v2.7.1 -> v2.7.1 2025-09-07T07:36:00.2849830Z * [new tag] v2.7.1-rc1 -> v2.7.1-rc1 2025-09-07T07:36:00.2850358Z * [new tag] v2.7.1-rc2 -> v2.7.1-rc2 2025-09-07T07:36:00.2850904Z * [new tag] v2.7.1-rc3 -> v2.7.1-rc3 2025-09-07T07:36:00.2851528Z * [new tag] v2.7.1-rc4 -> v2.7.1-rc4 2025-09-07T07:36:00.2852028Z * [new tag] v2.7.1-rc5 -> v2.7.1-rc5 2025-09-07T07:36:00.2852386Z * [new tag] v2.8.0 -> v2.8.0 2025-09-07T07:36:00.2852902Z * [new tag] v2.8.0-rc1 -> v2.8.0-rc1 2025-09-07T07:36:00.2853333Z * [new tag] v2.8.0-rc2 -> v2.8.0-rc2 2025-09-07T07:36:00.2853970Z * [new tag] v2.8.0-rc3 -> v2.8.0-rc3 2025-09-07T07:36:00.2854482Z * [new tag] v2.8.0-rc4 -> v2.8.0-rc4 2025-09-07T07:36:00.2854997Z * [new tag] v2.8.0-rc5 -> v2.8.0-rc5 2025-09-07T07:36:00.2855525Z * [new tag] v2.8.0-rc6 -> v2.8.0-rc6 2025-09-07T07:36:00.2855932Z * [new tag] v2.8.0-rc7 -> v2.8.0-rc7 2025-09-07T07:36:00.2856481Z * [new tag] v2.8.0-rc8 -> v2.8.0-rc8 2025-09-07T07:36:00.2856884Z * [new tag] whc_flight_1 -> whc_flight_1 2025-09-07T07:36:00.2857457Z * [new tag] whc_flight_2 -> whc_flight_2 2025-09-07T07:36:00.2857753Z * [new tag] whc_flight_4 -> whc_flight_4 2025-09-07T07:36:00.3279353Z [command]/usr/bin/git rev-parse --verify --quiet 93fb23d6fae7c4e82c4239a1033e522088742634^{object} 2025-09-07T07:36:00.3299978Z 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:36:00.3302605Z ##[endgroup] 2025-09-07T07:36:00.3302802Z ##[group]Determining the checkout info 2025-09-07T07:36:00.3303554Z ##[endgroup] 2025-09-07T07:36:00.3306134Z [command]/usr/bin/git sparse-checkout disable 2025-09-07T07:36:00.3339624Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-09-07T07:36:00.3358881Z ##[group]Checking out the ref 2025-09-07T07:36:00.3361814Z [command]/usr/bin/git checkout --progress --force 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:36:01.2963019Z Note: switching to '93fb23d6fae7c4e82c4239a1033e522088742634'. 2025-09-07T07:36:01.2963382Z 2025-09-07T07:36:01.2963640Z You are in 'detached HEAD' state. You can look around, make experimental 2025-09-07T07:36:01.2964054Z changes and commit them, and you can discard any commits you make in this 2025-09-07T07:36:01.2964460Z state without impacting any branches by switching back to a branch. 2025-09-07T07:36:01.2964691Z 2025-09-07T07:36:01.2965069Z If you want to create a new branch to retain commits you create, you may 2025-09-07T07:36:01.2965428Z do so (now or later) by using -c with the switch command. Example: 2025-09-07T07:36:01.2965642Z 2025-09-07T07:36:01.2965741Z git switch -c 2025-09-07T07:36:01.2965895Z 2025-09-07T07:36:01.2965988Z Or undo this operation with: 2025-09-07T07:36:01.2966119Z 2025-09-07T07:36:01.2966196Z git switch - 2025-09-07T07:36:01.2966299Z 2025-09-07T07:36:01.2966448Z Turn off this advice by setting config variable advice.detachedHead to false 2025-09-07T07:36:01.2966651Z 2025-09-07T07:36:01.2966767Z HEAD is now at 93fb23d6fae Build vLLM nightly wheels (#162000) 2025-09-07T07:36:01.3008497Z ##[endgroup] 2025-09-07T07:36:01.3008788Z ##[group]Setting up auth for fetching submodules 2025-09-07T07:36:01.3015139Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-09-07T07:36:01.3054784Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-09-07T07:36:01.3076324Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-09-07T07:36:01.3094650Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-09-07T07:36:01.3112611Z ##[endgroup] 2025-09-07T07:36:01.3112862Z ##[group]Fetching submodules 2025-09-07T07:36:01.3115269Z [command]/usr/bin/git submodule sync --recursive 2025-09-07T07:36:01.3374515Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2025-09-07T07:36:01.3630591Z Submodule 'android/libs/fbjni' (https://github.com/facebookincubator/fbjni.git) registered for path 'android/libs/fbjni' 2025-09-07T07:36:01.3631841Z Submodule 'third_party/NNPACK_deps/FP16' (https://github.com/Maratyszcza/FP16.git) registered for path 'third_party/FP16' 2025-09-07T07:36:01.3633523Z Submodule 'third_party/NNPACK_deps/FXdiv' (https://github.com/Maratyszcza/FXdiv.git) registered for path 'third_party/FXdiv' 2025-09-07T07:36:01.3635105Z Submodule 'third_party/NNPACK' (https://github.com/Maratyszcza/NNPACK.git) registered for path 'third_party/NNPACK' 2025-09-07T07:36:01.3968449Z Submodule 'third_party/NVTX' (https://github.com/NVIDIA/NVTX.git) registered for path 'third_party/NVTX' 2025-09-07T07:36:01.3969266Z Submodule 'third_party/VulkanMemoryAllocator' (https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git) registered for path 'third_party/VulkanMemoryAllocator' 2025-09-07T07:36:01.3973396Z Submodule 'third_party/XNNPACK' (https://github.com/google/XNNPACK.git) registered for path 'third_party/XNNPACK' 2025-09-07T07:36:01.3973964Z Submodule 'third_party/aiter' (https://github.com/ROCm/aiter.git) registered for path 'third_party/aiter' 2025-09-07T07:36:01.3975732Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/benchmark' 2025-09-07T07:36:01.3976379Z Submodule 'third_party/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/composable_kernel' 2025-09-07T07:36:01.3976970Z Submodule 'third_party/cpp-httplib' (https://github.com/yhirose/cpp-httplib.git) registered for path 'third_party/cpp-httplib' 2025-09-07T07:36:01.3978815Z Submodule 'third_party/cpuinfo' (https://github.com/pytorch/cpuinfo.git) registered for path 'third_party/cpuinfo' 2025-09-07T07:36:01.3980688Z Submodule 'third_party/cudnn_frontend' (https://github.com/NVIDIA/cudnn-frontend.git) registered for path 'third_party/cudnn_frontend' 2025-09-07T07:36:01.3993989Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/cutlass' 2025-09-07T07:36:01.3997650Z Submodule 'third_party/fbgemm' (https://github.com/pytorch/fbgemm) registered for path 'third_party/fbgemm' 2025-09-07T07:36:01.3999818Z Submodule 'third_party/flash-attention' (https://github.com/Dao-AILab/flash-attention.git) registered for path 'third_party/flash-attention' 2025-09-07T07:36:01.4002032Z Submodule 'third_party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'third_party/flatbuffers' 2025-09-07T07:36:01.4003808Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/fmt' 2025-09-07T07:36:01.4006039Z Submodule 'third_party/gemmlowp/gemmlowp' (https://github.com/google/gemmlowp.git) registered for path 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:36:01.4008141Z Submodule 'third_party/gloo' (https://github.com/pytorch/gloo) registered for path 'third_party/gloo' 2025-09-07T07:36:01.4011373Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/googletest' 2025-09-07T07:36:01.4013562Z Submodule 'third_party/ideep' (https://github.com/intel/ideep) registered for path 'third_party/ideep' 2025-09-07T07:36:01.4015823Z Submodule 'third_party/ittapi' (https://github.com/intel/ittapi.git) registered for path 'third_party/ittapi' 2025-09-07T07:36:01.4029795Z Submodule 'third_party/kineto' (https://github.com/pytorch/kineto) registered for path 'third_party/kineto' 2025-09-07T07:36:01.4032132Z Submodule 'third_party/kleidiai' (https://github.com/ARM-software/kleidiai.git) registered for path 'third_party/kleidiai' 2025-09-07T07:36:01.4034429Z Submodule 'third_party/mimalloc' (https://github.com/microsoft/mimalloc.git) registered for path 'third_party/mimalloc' 2025-09-07T07:36:01.4036795Z Submodule 'third_party/nlohmann' (https://github.com/nlohmann/json.git) registered for path 'third_party/nlohmann' 2025-09-07T07:36:01.4039185Z Submodule 'third_party/onnx' (https://github.com/onnx/onnx.git) registered for path 'third_party/onnx' 2025-09-07T07:36:01.4041704Z Submodule 'third_party/opentelemetry-cpp' (https://github.com/open-telemetry/opentelemetry-cpp.git) registered for path 'third_party/opentelemetry-cpp' 2025-09-07T07:36:01.4044395Z Submodule 'third_party/pocketfft' (https://github.com/mreineck/pocketfft) registered for path 'third_party/pocketfft' 2025-09-07T07:36:01.4047032Z Submodule 'third_party/protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'third_party/protobuf' 2025-09-07T07:36:01.4049625Z Submodule 'third_party/NNPACK_deps/psimd' (https://github.com/Maratyszcza/psimd.git) registered for path 'third_party/psimd' 2025-09-07T07:36:01.4064999Z Submodule 'third_party/NNPACK_deps/pthreadpool' (https://github.com/Maratyszcza/pthreadpool.git) registered for path 'third_party/pthreadpool' 2025-09-07T07:36:01.4067623Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/pybind11' 2025-09-07T07:36:01.4070549Z Submodule 'third_party/python-peachpy' (https://github.com/malfet/PeachPy.git) registered for path 'third_party/python-peachpy' 2025-09-07T07:36:01.4073350Z Submodule 'third_party/sleef' (https://github.com/shibatch/sleef) registered for path 'third_party/sleef' 2025-09-07T07:36:01.4076237Z Submodule 'third_party/tensorpipe' (https://github.com/pytorch/tensorpipe.git) registered for path 'third_party/tensorpipe' 2025-09-07T07:36:01.4100026Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/android/libs/fbjni'... 2025-09-07T07:36:01.6421697Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FXdiv'... 2025-09-07T07:36:01.6422546Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FP16'... 2025-09-07T07:36:01.6423304Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/psimd'... 2025-09-07T07:36:01.6424085Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NNPACK'... 2025-09-07T07:36:01.6424888Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pthreadpool'... 2025-09-07T07:36:01.6425696Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pocketfft'... 2025-09-07T07:36:01.6838137Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NVTX'... 2025-09-07T07:36:01.7984940Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/python-peachpy'... 2025-09-07T07:36:01.7985871Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep'... 2025-09-07T07:36:01.7986677Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gloo'... 2025-09-07T07:36:01.7987475Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/benchmark'... 2025-09-07T07:36:01.7988383Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/VulkanMemoryAllocator'... 2025-09-07T07:36:02.6483160Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gemmlowp/gemmlowp'... 2025-09-07T07:36:02.6483937Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe'... 2025-09-07T07:36:02.6484663Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ittapi'... 2025-09-07T07:36:02.6485171Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kleidiai'... 2025-09-07T07:36:02.6485670Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention'... 2025-09-07T07:36:02.6486195Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpp-httplib'... 2025-09-07T07:36:02.6486839Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpuinfo'... 2025-09-07T07:36:02.6487275Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/sleef'... 2025-09-07T07:36:02.6487761Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/mimalloc'... 2025-09-07T07:36:02.6488226Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/googletest'... 2025-09-07T07:36:02.6488718Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pybind11'... 2025-09-07T07:36:02.6489391Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fmt'... 2025-09-07T07:36:02.6489869Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cudnn_frontend'... 2025-09-07T07:36:02.6490318Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto'... 2025-09-07T07:36:02.7484340Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/XNNPACK'... 2025-09-07T07:36:10.6391623Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flatbuffers'... 2025-09-07T07:36:10.6392110Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm'... 2025-09-07T07:36:10.6392505Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cutlass'... 2025-09-07T07:36:10.6392885Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx'... 2025-09-07T07:36:10.6393303Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/composable_kernel'... 2025-09-07T07:36:10.6393761Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter'... 2025-09-07T07:36:10.6394184Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp'... 2025-09-07T07:36:10.6394598Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/nlohmann'... 2025-09-07T07:36:10.6394978Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf'... 2025-09-07T07:36:10.6511895Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2025-09-07T07:36:10.6601784Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-09-07T07:36:10.6673447Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-09-07T07:36:10.6858937Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2025-09-07T07:36:10.7470674Z Submodule path 'third_party/NVTX': checked out '2942f167cc30c5e3a44a2aecd5b0d9c07ff61a07' 2025-09-07T07:36:10.7847519Z Submodule path 'third_party/VulkanMemoryAllocator': checked out '1d8f600fd424278486eade7ed3e877c99f0846b1' 2025-09-07T07:36:11.2757466Z Submodule path 'third_party/XNNPACK': checked out '51a0103656eff6fc9bfd39a4597923c4b542c883' 2025-09-07T07:36:11.3932145Z Submodule path 'third_party/aiter': checked out '01aae101b9e5e94d6c16a9514c9fb8df99c93150' 2025-09-07T07:36:11.3942579Z Submodule '3rdparty/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:36:11.3965765Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter/3rdparty/composable_kernel'... 2025-09-07T07:36:14.4550056Z Submodule path 'third_party/aiter/3rdparty/composable_kernel': checked out 'cffe8fa2a442ac8e80dd236a1a5d24fe3d7e0cbf' 2025-09-07T07:36:14.4730066Z Submodule path 'third_party/benchmark': checked out '299e5928955cc62af9968370293b916f5130916f' 2025-09-07T07:36:14.7017149Z Submodule path 'third_party/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-09-07T07:36:14.7363371Z Submodule path 'third_party/cpp-httplib': checked out '89c932f313c6437c38f2982869beacc89c2f2246' 2025-09-07T07:36:14.8135490Z Submodule path 'third_party/cpuinfo': checked out '5e3d2445e6a84d9599bee2bf78edbb4d80865e1d' 2025-09-07T07:36:14.8479914Z Submodule path 'third_party/cudnn_frontend': checked out 'f937055efc6d414d11f4c6577e3977fe74f35fb6' 2025-09-07T07:36:15.3313496Z Submodule path 'third_party/cutlass': checked out 'e51efbfe18fe4f4cbb66ab814c55bf4aa0185491' 2025-09-07T07:36:15.4397518Z Submodule path 'third_party/fbgemm': checked out '4b39c551efe15e6bbade20565b0ceb2d8ce3352d' 2025-09-07T07:36:15.4407968Z Submodule 'external/asmjit' (https://github.com/asmjit/asmjit.git) registered for path 'third_party/fbgemm/external/asmjit' 2025-09-07T07:36:15.4409265Z Submodule 'external/composable_kernel' (https://github.com/jwfromm/composable_kernel.git) registered for path 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:36:15.4410388Z Submodule 'external/cpuinfo' (https://github.com/pytorch/cpuinfo) registered for path 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:36:15.4411959Z Submodule 'external/cutlass' (https://github.com/jwfromm/cutlass) registered for path 'third_party/fbgemm/external/cutlass' 2025-09-07T07:36:15.4413267Z Submodule 'external/googletest' (https://github.com/google/googletest) registered for path 'third_party/fbgemm/external/googletest' 2025-09-07T07:36:15.4414741Z Submodule 'external/hipify_torch' (https://github.com/ROCmSoftwarePlatform/hipify_torch.git) registered for path 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:36:15.4416050Z Submodule 'external/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/fbgemm/external/json' 2025-09-07T07:36:15.4436638Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/asmjit'... 2025-09-07T07:36:16.4572787Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/hipify_torch'... 2025-09-07T07:36:16.4573499Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cpuinfo'... 2025-09-07T07:36:16.4574092Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/googletest'... 2025-09-07T07:36:16.4574689Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/composable_kernel'... 2025-09-07T07:36:16.5571341Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cutlass'... 2025-09-07T07:36:17.1868714Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/json'... 2025-09-07T07:36:20.7050369Z Submodule path 'third_party/fbgemm/external/asmjit': checked out 'a3199e8857792cd10b7589ff5d58343d2c9008ea' 2025-09-07T07:36:20.8860262Z Submodule path 'third_party/fbgemm/external/composable_kernel': checked out 'b1281b8b08d973a7064f864f47eeb30f3e2596e9' 2025-09-07T07:36:20.9650238Z Submodule path 'third_party/fbgemm/external/cpuinfo': checked out '6543fec09b2f04ac4a666882998b534afc9c1349' 2025-09-07T07:36:21.4418202Z Submodule path 'third_party/fbgemm/external/cutlass': checked out '311f3c8e51dc0eb56310cfc6980bf63d0fbd7917' 2025-09-07T07:36:21.4778979Z Submodule path 'third_party/fbgemm/external/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-09-07T07:36:21.4866767Z Submodule path 'third_party/fbgemm/external/hipify_torch': checked out '63b6a7b541fa7f08f8475ca7d74054db36ff2691' 2025-09-07T07:36:21.5700980Z Submodule path 'third_party/fbgemm/external/json': checked out '9cca280a4d0ccf0c08f47a99aa71d1b0e52f8d03' 2025-09-07T07:36:21.6212840Z Submodule path 'third_party/flash-attention': checked out '979702c87a8713a8e0a5e9fee122b90d2ef13be5' 2025-09-07T07:36:21.6227752Z Submodule 'csrc/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:36:21.6228535Z Submodule 'csrc/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:36:21.6250172Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/composable_kernel'... 2025-09-07T07:36:24.4286633Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/cutlass'... 2025-09-07T07:36:24.5941847Z Submodule path 'third_party/flash-attention/csrc/composable_kernel': checked out '888317e698e9803c62bd38568abc9e05d7709f33' 2025-09-07T07:36:25.0199249Z Submodule path 'third_party/flash-attention/csrc/cutlass': checked out 'c506e16788cb08416a4a57e11a9067beeee29420' 2025-09-07T07:36:25.1171900Z Submodule path 'third_party/flatbuffers': checked out 'a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757' 2025-09-07T07:36:25.1423733Z Submodule path 'third_party/fmt': checked out '40626af88bd7df9a5fb80be7b25ac85b122d6c21' 2025-09-07T07:36:25.1723084Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2025-09-07T07:36:25.1904801Z Submodule path 'third_party/gloo': checked out 'c7b7b022c124d9643957d9bd55f57ac59fce8fa2' 2025-09-07T07:36:25.2243699Z Submodule path 'third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-09-07T07:36:25.2340439Z Submodule path 'third_party/ideep': checked out '719d8e6cd7f7a0e01b155657526d693acf97c2b3' 2025-09-07T07:36:25.2352351Z Submodule 'mkl-dnn' (https://github.com/intel/mkl-dnn.git) registered for path 'third_party/ideep/mkl-dnn' 2025-09-07T07:36:25.2373898Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep/mkl-dnn'... 2025-09-07T07:36:36.0893166Z Submodule path 'third_party/ideep/mkl-dnn': checked out '8d263e693366ef8db40acc569cc7d8edf644556d' 2025-09-07T07:36:36.1046496Z Submodule path 'third_party/ittapi': checked out 'dec1d23ca65ab069d225dfe40dea14f455170959' 2025-09-07T07:36:36.1811059Z Submodule path 'third_party/kineto': checked out '5e7501833f1021ce6f618572d3baf657b6319658' 2025-09-07T07:36:36.1821598Z Submodule 'libkineto/third_party/dynolog' (https://github.com/facebookincubator/dynolog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:36:36.1822544Z Submodule 'libkineto/third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:36:36.1823801Z Submodule 'libkineto/third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:36:36.1845073Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog'... 2025-09-07T07:36:36.7769298Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/fmt'... 2025-09-07T07:36:37.0115957Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/googletest'... 2025-09-07T07:36:37.0762755Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out '7d04a0053a845370ae06ce317a22a48e9edcc74e' 2025-09-07T07:36:37.0774839Z Submodule 'third_party/DCGM' (https://github.com/NVIDIA/DCGM.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:36:37.0791489Z Submodule 'third_party/cpr' (https://github.com/libcpr/cpr.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:36:37.0792191Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:36:37.0792865Z Submodule 'third_party/gflags' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:36:37.0793564Z Submodule 'third_party/glog' (https://github.com/google/glog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:36:37.0794269Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:36:37.0794973Z Submodule 'third_party/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:36:37.0795616Z Submodule 'third_party/pfs' (https://github.com/dtrugman/pfs.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:36:37.0805325Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM'... 2025-09-07T07:36:38.2362976Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/pfs'... 2025-09-07T07:36:38.2363917Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags'... 2025-09-07T07:36:38.2364530Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/cpr'... 2025-09-07T07:36:38.2365133Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/glog'... 2025-09-07T07:36:38.2365756Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/googletest'... 2025-09-07T07:36:38.2366363Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/fmt'... 2025-09-07T07:36:38.3366073Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/json'... 2025-09-07T07:36:43.6778576Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2025-09-07T07:36:43.6912118Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2025-09-07T07:36:43.7191428Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2025-09-07T07:36:43.7295665Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2025-09-07T07:36:43.7307533Z Submodule 'doc' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:36:43.7327240Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc'... 2025-09-07T07:36:43.9800833Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2025-09-07T07:36:43.9943892Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2025-09-07T07:36:44.0265001Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '58d77fa8070e8cec2dc1ed015d66b454c8d78850' 2025-09-07T07:36:44.1045933Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2025-09-07T07:36:44.1172618Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2025-09-07T07:36:44.1464818Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out '0041a40c1350ba702d475b9c4ad62da77caea164' 2025-09-07T07:36:44.1912120Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '7aca84427f224eeed3144123d5230d5871e93347' 2025-09-07T07:36:44.2240351Z Submodule path 'third_party/kleidiai': checked out 'cca02c2f69dd18e1f12647c1c0bdc8cf90e680c7' 2025-09-07T07:36:44.2536271Z Submodule path 'third_party/mimalloc': checked out 'fbd8b99c2b828428947d70fdc046bb55609be93e' 2025-09-07T07:36:44.3407018Z Submodule path 'third_party/nlohmann': checked out '55f93686c01528224f448c19128836e7df245f72' 2025-09-07T07:36:44.5979991Z Submodule path 'third_party/onnx': checked out 'e709452ef2bbc1d113faf678c24e6d3467696e83' 2025-09-07T07:36:44.6001208Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/onnx/third_party/pybind11' 2025-09-07T07:36:44.6020173Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx/third_party/pybind11'... 2025-09-07T07:36:45.4120280Z Submodule path 'third_party/onnx/third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-09-07T07:36:45.4597289Z Submodule path 'third_party/opentelemetry-cpp': checked out 'a799f4aed9c94b765dcdaabaeab7d5e7e2310878' 2025-09-07T07:36:45.4607511Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark) registered for path 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:36:45.4608716Z Submodule 'third_party/googletest' (https://github.com/google/googletest) registered for path 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:36:45.4609951Z Submodule 'third_party/ms-gsl' (https://github.com/microsoft/GSL) registered for path 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:36:45.4611248Z Submodule 'third_party/nlohmann-json' (https://github.com/nlohmann/json) registered for path 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:36:45.4612916Z Submodule 'third_party/opentelemetry-proto' (https://github.com/open-telemetry/opentelemetry-proto) registered for path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:36:45.4614212Z Submodule 'third_party/opentracing-cpp' (https://github.com/opentracing/opentracing-cpp.git) registered for path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:36:45.4615618Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:36:45.4617009Z Submodule 'tools/vcpkg' (https://github.com/Microsoft/vcpkg) registered for path 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:36:45.4639549Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/benchmark'... 2025-09-07T07:36:45.8104675Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentracing-cpp'... 2025-09-07T07:36:45.8105461Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentelemetry-proto'... 2025-09-07T07:36:45.8106348Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/ms-gsl'... 2025-09-07T07:36:45.8106981Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp'... 2025-09-07T07:36:45.9105263Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/googletest'... 2025-09-07T07:36:46.3301934Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/nlohmann-json'... 2025-09-07T07:36:51.3401507Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/tools/vcpkg'... 2025-09-07T07:36:51.5932991Z Submodule path 'third_party/opentelemetry-cpp/third_party/benchmark': checked out 'd572f4777349d43653b21d6c2fc63020ab326db2' 2025-09-07T07:36:51.6248570Z Submodule path 'third_party/opentelemetry-cpp/third_party/googletest': checked out 'b796f7d44681514f58a683a3a71ff17c94edb0c1' 2025-09-07T07:36:51.6378839Z Submodule path 'third_party/opentelemetry-cpp/third_party/ms-gsl': checked out '6f4529395c5b7c2d661812257cd6780c67e54afa' 2025-09-07T07:36:51.7187826Z Submodule path 'third_party/opentelemetry-cpp/third_party/nlohmann-json': checked out 'bc889afb4c5bf1c0d8ee29ef35eaaf4c8bef8a5d' 2025-09-07T07:36:51.7290612Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto': checked out '4ca4f0335c63cda7ab31ea7ed70d6553aee14dce' 2025-09-07T07:36:51.7397268Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp': checked out '06b57f48ded1fa3bdd3d4346f6ef29e40e08eaf5' 2025-09-07T07:36:51.7510139Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp': checked out 'c9ffcdda9086ffd9e1283ea7a0276d831f3c8a8d' 2025-09-07T07:36:51.7523111Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:36:51.7524396Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:36:51.7546030Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-09-07T07:36:53.1005847Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest'... 2025-09-07T07:36:53.2980046Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'eefb26f82b233268fc98577d265352720d477ba4' 2025-09-07T07:36:53.3330360Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-09-07T07:36:53.6310666Z Submodule path 'third_party/opentelemetry-cpp/tools/vcpkg': checked out '8eb57355a4ffb410a2e94c07b4dca2dffbee8e50' 2025-09-07T07:36:53.6399351Z Submodule path 'third_party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-09-07T07:36:53.8530163Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2025-09-07T07:36:53.8543674Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:36:53.8544985Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/protobuf/third_party/googletest' 2025-09-07T07:36:53.8567019Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/benchmark'... 2025-09-07T07:36:54.4727655Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/googletest'... 2025-09-07T07:36:54.8197691Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2025-09-07T07:36:54.8747665Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2025-09-07T07:36:54.8818871Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2025-09-07T07:36:54.8916336Z Submodule path 'third_party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-09-07T07:36:54.9223328Z Submodule path 'third_party/pybind11': checked out 'f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8' 2025-09-07T07:36:54.9436160Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2025-09-07T07:36:54.9783851Z Submodule path 'third_party/sleef': checked out '5a1d179df9cf652951b59010a2d2075372d67f68' 2025-09-07T07:36:54.9984498Z Submodule path 'third_party/tensorpipe': checked out 'af0118d13e52f5a08841464a768e01a0bf3e3075' 2025-09-07T07:36:54.9995630Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:36:54.9996789Z Submodule 'third_party/libnop' (https://github.com/google/libnop.git) registered for path 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:36:54.9997998Z Submodule 'third_party/libuv' (https://github.com/libuv/libuv.git) registered for path 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:36:54.9999356Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:36:55.0022466Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/googletest'... 2025-09-07T07:36:56.1074198Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libnop'... 2025-09-07T07:36:56.1224502Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libuv'... 2025-09-07T07:36:56.3320644Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11'... 2025-09-07T07:36:56.3773166Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2025-09-07T07:36:56.3893122Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2025-09-07T07:36:56.4449390Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '5152db2cbfeb5582e9c27c5ea1dba2cd9e10759b' 2025-09-07T07:36:56.4668018Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2025-09-07T07:36:56.4677522Z Submodule 'tools/clang' (https://github.com/wjakob/clang-cindex-python3) registered for path 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:36:56.4698935Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11/tools/clang'... 2025-09-07T07:36:56.7472326Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2025-09-07T07:36:56.7502795Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-09-07T07:36:56.7759657Z Entering 'android/libs/fbjni' 2025-09-07T07:36:56.7791497Z Entering 'third_party/FP16' 2025-09-07T07:36:56.7826298Z Entering 'third_party/FXdiv' 2025-09-07T07:36:56.7860054Z Entering 'third_party/NNPACK' 2025-09-07T07:36:56.7892255Z Entering 'third_party/NVTX' 2025-09-07T07:36:56.7926468Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:36:56.7959841Z Entering 'third_party/XNNPACK' 2025-09-07T07:36:56.8004300Z Entering 'third_party/aiter' 2025-09-07T07:36:56.8041014Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:36:56.8079331Z Entering 'third_party/benchmark' 2025-09-07T07:36:56.8107842Z Entering 'third_party/composable_kernel' 2025-09-07T07:36:56.8146790Z Entering 'third_party/cpp-httplib' 2025-09-07T07:36:56.8180419Z Entering 'third_party/cpuinfo' 2025-09-07T07:36:56.8213948Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:36:56.8249236Z Entering 'third_party/cutlass' 2025-09-07T07:36:56.8291490Z Entering 'third_party/fbgemm' 2025-09-07T07:36:56.8324999Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:36:56.8356473Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:36:56.8396271Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:36:56.8428774Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:36:56.8468158Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:36:56.8497968Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:36:56.8531590Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:36:56.8569667Z Entering 'third_party/flash-attention' 2025-09-07T07:36:56.8602890Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:36:56.8639032Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:36:56.8677445Z Entering 'third_party/flatbuffers' 2025-09-07T07:36:56.8711334Z Entering 'third_party/fmt' 2025-09-07T07:36:56.8744669Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:36:56.8811766Z Entering 'third_party/gloo' 2025-09-07T07:36:56.8842021Z Entering 'third_party/googletest' 2025-09-07T07:36:56.8875546Z Entering 'third_party/ideep' 2025-09-07T07:36:56.8909338Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:36:56.8943776Z Entering 'third_party/ittapi' 2025-09-07T07:36:56.8979242Z Entering 'third_party/kineto' 2025-09-07T07:36:56.9012365Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:36:56.9046504Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:36:56.9079492Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:36:56.9110132Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:36:56.9141606Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:36:56.9173742Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:36:56.9207299Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:36:56.9238884Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:36:56.9272290Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:36:56.9304198Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:36:56.9339130Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:36:56.9371967Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:36:56.9405228Z Entering 'third_party/kleidiai' 2025-09-07T07:36:56.9437878Z Entering 'third_party/mimalloc' 2025-09-07T07:36:56.9469933Z Entering 'third_party/nlohmann' 2025-09-07T07:36:56.9502735Z Entering 'third_party/onnx' 2025-09-07T07:36:56.9550189Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:36:56.9582800Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:36:56.9616920Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:36:56.9651015Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:36:56.9680725Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:36:56.9713398Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:36:56.9746644Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:36:56.9779848Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:36:56.9811323Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:36:56.9844966Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:36:56.9880252Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:36:56.9909881Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:36:56.9956272Z Entering 'third_party/pocketfft' 2025-09-07T07:36:56.9992027Z Entering 'third_party/protobuf' 2025-09-07T07:36:57.0027977Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:36:57.0059932Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:36:57.0096492Z Entering 'third_party/psimd' 2025-09-07T07:36:57.0127919Z Entering 'third_party/pthreadpool' 2025-09-07T07:36:57.0160461Z Entering 'third_party/pybind11' 2025-09-07T07:36:57.0191461Z Entering 'third_party/python-peachpy' 2025-09-07T07:36:57.0223756Z Entering 'third_party/sleef' 2025-09-07T07:36:57.0255526Z Entering 'third_party/tensorpipe' 2025-09-07T07:36:57.0288193Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:36:57.0319117Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:36:57.0350447Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:36:57.0384164Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:36:57.0416214Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:36:57.0464366Z ##[endgroup] 2025-09-07T07:36:57.0466453Z ##[group]Persisting credentials for submodules 2025-09-07T07:36:57.0474042Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-09-07T07:36:57.0718009Z Entering 'android/libs/fbjni' 2025-09-07T07:36:57.0766731Z Entering 'third_party/FP16' 2025-09-07T07:36:57.0810244Z Entering 'third_party/FXdiv' 2025-09-07T07:36:57.0856146Z Entering 'third_party/NNPACK' 2025-09-07T07:36:57.0901013Z Entering 'third_party/NVTX' 2025-09-07T07:36:57.0943921Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:36:57.0988763Z Entering 'third_party/XNNPACK' 2025-09-07T07:36:57.1040260Z Entering 'third_party/aiter' 2025-09-07T07:36:57.1081604Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:36:57.1132830Z Entering 'third_party/benchmark' 2025-09-07T07:36:57.1177091Z Entering 'third_party/composable_kernel' 2025-09-07T07:36:57.1225232Z Entering 'third_party/cpp-httplib' 2025-09-07T07:36:57.1272516Z Entering 'third_party/cpuinfo' 2025-09-07T07:36:57.1317391Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:36:57.1364598Z Entering 'third_party/cutlass' 2025-09-07T07:36:57.1416069Z Entering 'third_party/fbgemm' 2025-09-07T07:36:57.1463157Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:36:57.1505376Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:36:57.1552272Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:36:57.1594312Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:36:57.1641697Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:36:57.1685099Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:36:57.1727309Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:36:57.1777311Z Entering 'third_party/flash-attention' 2025-09-07T07:36:57.1818089Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:36:57.1863683Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:36:57.1916845Z Entering 'third_party/flatbuffers' 2025-09-07T07:36:57.1962948Z Entering 'third_party/fmt' 2025-09-07T07:36:57.2004423Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:36:57.2048318Z Entering 'third_party/gloo' 2025-09-07T07:36:57.2091114Z Entering 'third_party/googletest' 2025-09-07T07:36:57.2135006Z Entering 'third_party/ideep' 2025-09-07T07:36:57.2183513Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:36:57.2229559Z Entering 'third_party/ittapi' 2025-09-07T07:36:57.2273264Z Entering 'third_party/kineto' 2025-09-07T07:36:57.2317360Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:36:57.2361452Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:36:57.2404914Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:36:57.2445605Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:36:57.2491004Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:36:57.2532467Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:36:57.2578915Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:36:57.2618134Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:36:57.2659718Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:36:57.2705034Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:36:57.2752006Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:36:57.2795956Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:36:57.2839203Z Entering 'third_party/kleidiai' 2025-09-07T07:36:57.2880309Z Entering 'third_party/mimalloc' 2025-09-07T07:36:57.2920514Z Entering 'third_party/nlohmann' 2025-09-07T07:36:57.2968118Z Entering 'third_party/onnx' 2025-09-07T07:36:57.3023684Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:36:57.3069389Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:36:57.3113897Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:36:57.3154715Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:36:57.3197028Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:36:57.3236006Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:36:57.3280650Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:36:57.3326291Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:36:57.3369467Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:36:57.3411471Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:36:57.3456941Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:36:57.3500448Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:36:57.3560545Z Entering 'third_party/pocketfft' 2025-09-07T07:36:57.3603657Z Entering 'third_party/protobuf' 2025-09-07T07:36:57.3647444Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:36:57.3688282Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:36:57.3733820Z Entering 'third_party/psimd' 2025-09-07T07:36:57.3778141Z Entering 'third_party/pthreadpool' 2025-09-07T07:36:57.3823740Z Entering 'third_party/pybind11' 2025-09-07T07:36:57.3871255Z Entering 'third_party/python-peachpy' 2025-09-07T07:36:57.3911385Z Entering 'third_party/sleef' 2025-09-07T07:36:57.3954201Z Entering 'third_party/tensorpipe' 2025-09-07T07:36:57.3998404Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:36:57.4041097Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:36:57.4082418Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:36:57.4127130Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:36:57.4170572Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:36:57.4231309Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-09-07T07:36:57.4470431Z Entering 'android/libs/fbjni' 2025-09-07T07:36:57.4507366Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-09-07T07:36:57.4517001Z Entering 'third_party/FP16' 2025-09-07T07:36:57.4556623Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-09-07T07:36:57.4568353Z Entering 'third_party/FXdiv' 2025-09-07T07:36:57.4610352Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-09-07T07:36:57.4619938Z Entering 'third_party/NNPACK' 2025-09-07T07:36:57.4657974Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-09-07T07:36:57.4672233Z Entering 'third_party/NVTX' 2025-09-07T07:36:57.4710350Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-09-07T07:36:57.4724448Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:36:57.4766154Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-09-07T07:36:57.4780521Z Entering 'third_party/XNNPACK' 2025-09-07T07:36:57.4820583Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-09-07T07:36:57.4843972Z Entering 'third_party/aiter' 2025-09-07T07:36:57.4882747Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-09-07T07:36:57.4895177Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:36:57.4934327Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-09-07T07:36:57.4952967Z Entering 'third_party/benchmark' 2025-09-07T07:36:57.4993783Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-09-07T07:36:57.5008167Z Entering 'third_party/composable_kernel' 2025-09-07T07:36:57.5050257Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-09-07T07:36:57.5069480Z Entering 'third_party/cpp-httplib' 2025-09-07T07:36:57.5108040Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-09-07T07:36:57.5120129Z Entering 'third_party/cpuinfo' 2025-09-07T07:36:57.5159986Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-09-07T07:36:57.5174627Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:36:57.5215983Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-09-07T07:36:57.5229384Z Entering 'third_party/cutlass' 2025-09-07T07:36:57.5268583Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-09-07T07:36:57.5285682Z Entering 'third_party/fbgemm' 2025-09-07T07:36:57.5324120Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-09-07T07:36:57.5339578Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:36:57.5381133Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-09-07T07:36:57.5394512Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:36:57.5435019Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-09-07T07:36:57.5452410Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:36:57.5492697Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-09-07T07:36:57.5506553Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:36:57.5546848Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-09-07T07:36:57.5564948Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:36:57.5605666Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-09-07T07:36:57.5619696Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:36:57.5660548Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-09-07T07:36:57.5672513Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:36:57.5711163Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-09-07T07:36:57.5728236Z Entering 'third_party/flash-attention' 2025-09-07T07:36:57.5767757Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-09-07T07:36:57.5779255Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:36:57.5816309Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-09-07T07:36:57.5832392Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:36:57.5870324Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-09-07T07:36:57.5889792Z Entering 'third_party/flatbuffers' 2025-09-07T07:36:57.5927421Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-09-07T07:36:57.5941193Z Entering 'third_party/fmt' 2025-09-07T07:36:57.5984499Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-09-07T07:36:57.5996979Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:36:57.6037716Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-09-07T07:36:57.6049809Z Entering 'third_party/gloo' 2025-09-07T07:36:57.6089370Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-09-07T07:36:57.6102387Z Entering 'third_party/googletest' 2025-09-07T07:36:57.6139497Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:36:57.6151873Z Entering 'third_party/ideep' 2025-09-07T07:36:57.6194078Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-09-07T07:36:57.6207150Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:36:57.6247272Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-09-07T07:36:57.6264951Z Entering 'third_party/ittapi' 2025-09-07T07:36:57.6305216Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-09-07T07:36:57.6316280Z Entering 'third_party/kineto' 2025-09-07T07:36:57.6357005Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-09-07T07:36:57.6368896Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:36:57.6409192Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-09-07T07:36:57.6421619Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:36:57.6464087Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-09-07T07:36:57.6477172Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:36:57.6515971Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-09-07T07:36:57.6528763Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:36:57.6568665Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-09-07T07:36:57.6581493Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:36:57.6620518Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-09-07T07:36:57.6631836Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:36:57.6673931Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-09-07T07:36:57.6688502Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:36:57.6727291Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-09-07T07:36:57.6740190Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:36:57.6780097Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:36:57.6791756Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:36:57.6829324Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-09-07T07:36:57.6840881Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:36:57.6882060Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-09-07T07:36:57.6894590Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:36:57.6934080Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-09-07T07:36:57.6947573Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:36:57.6988267Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-09-07T07:36:57.7002226Z Entering 'third_party/kleidiai' 2025-09-07T07:36:57.7043308Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-09-07T07:36:57.7056297Z Entering 'third_party/mimalloc' 2025-09-07T07:36:57.7096185Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-09-07T07:36:57.7107167Z Entering 'third_party/nlohmann' 2025-09-07T07:36:57.7146657Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-09-07T07:36:57.7161585Z Entering 'third_party/onnx' 2025-09-07T07:36:57.7204364Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-09-07T07:36:57.7230182Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:36:57.7271315Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-09-07T07:36:57.7286198Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:36:57.7323217Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-09-07T07:36:57.7336908Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:36:57.7374013Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-09-07T07:36:57.7386139Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:36:57.7424539Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:36:57.7437504Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:36:57.7478431Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-09-07T07:36:57.7491831Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:36:57.7530161Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-09-07T07:36:57.7541726Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:36:57.7582316Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-09-07T07:36:57.7593956Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:36:57.7631474Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-09-07T07:36:57.7644741Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:36:57.7686249Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-09-07T07:36:57.7697242Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:36:57.7737647Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-09-07T07:36:57.7752371Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:36:57.7792387Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-09-07T07:36:57.7804071Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:36:57.7840771Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-09-07T07:36:57.7869278Z Entering 'third_party/pocketfft' 2025-09-07T07:36:57.7910023Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-09-07T07:36:57.7923274Z Entering 'third_party/protobuf' 2025-09-07T07:36:57.7962987Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-09-07T07:36:57.7979401Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:36:57.8019057Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-09-07T07:36:57.8032070Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:36:57.8073602Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:36:57.8087501Z Entering 'third_party/psimd' 2025-09-07T07:36:57.8127648Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-09-07T07:36:57.8140459Z Entering 'third_party/pthreadpool' 2025-09-07T07:36:57.8180367Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-09-07T07:36:57.8191984Z Entering 'third_party/pybind11' 2025-09-07T07:36:57.8231169Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-09-07T07:36:57.8243437Z Entering 'third_party/python-peachpy' 2025-09-07T07:36:57.8286368Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-09-07T07:36:57.8296942Z Entering 'third_party/sleef' 2025-09-07T07:36:57.8333449Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-09-07T07:36:57.8344441Z Entering 'third_party/tensorpipe' 2025-09-07T07:36:57.8385798Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-09-07T07:36:57.8396847Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:36:57.8435310Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:36:57.8446963Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:36:57.8486278Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-09-07T07:36:57.8497034Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:36:57.8536803Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-09-07T07:36:57.8548703Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:36:57.8585476Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-09-07T07:36:57.8597949Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:36:57.8634582Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-09-07T07:36:57.9546775Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-09-07T07:36:57.9787300Z Entering 'android/libs/fbjni' 2025-09-07T07:36:57.9817893Z Entering 'third_party/FP16' 2025-09-07T07:36:57.9850953Z Entering 'third_party/FXdiv' 2025-09-07T07:36:57.9882243Z Entering 'third_party/NNPACK' 2025-09-07T07:36:57.9913675Z Entering 'third_party/NVTX' 2025-09-07T07:36:57.9947241Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:36:57.9978759Z Entering 'third_party/XNNPACK' 2025-09-07T07:36:58.0020424Z Entering 'third_party/aiter' 2025-09-07T07:36:58.0052750Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:36:58.0089885Z Entering 'third_party/benchmark' 2025-09-07T07:36:58.0118416Z Entering 'third_party/composable_kernel' 2025-09-07T07:36:58.0156949Z Entering 'third_party/cpp-httplib' 2025-09-07T07:36:58.0189174Z Entering 'third_party/cpuinfo' 2025-09-07T07:36:58.0223226Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:36:58.0258091Z Entering 'third_party/cutlass' 2025-09-07T07:36:58.0296203Z Entering 'third_party/fbgemm' 2025-09-07T07:36:58.0328251Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:36:58.0360150Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:36:58.0397343Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:36:58.0432023Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:36:58.0473907Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:36:58.0502208Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:36:58.0530600Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:36:58.0564565Z Entering 'third_party/flash-attention' 2025-09-07T07:36:58.0598085Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:36:58.0636951Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:36:58.0676156Z Entering 'third_party/flatbuffers' 2025-09-07T07:36:58.0707498Z Entering 'third_party/fmt' 2025-09-07T07:36:58.0740330Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:36:58.0774648Z Entering 'third_party/gloo' 2025-09-07T07:36:58.0809570Z Entering 'third_party/googletest' 2025-09-07T07:36:58.0843273Z Entering 'third_party/ideep' 2025-09-07T07:36:58.0876831Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:36:58.0913031Z Entering 'third_party/ittapi' 2025-09-07T07:36:58.0945776Z Entering 'third_party/kineto' 2025-09-07T07:36:58.0978277Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:36:58.1009204Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:36:58.1041541Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:36:58.1073510Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:36:58.1105604Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:36:58.1134931Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:36:58.1168637Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:36:58.1200230Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:36:58.1230694Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:36:58.1263114Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:36:58.1295545Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:36:58.1325388Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:36:58.1360549Z Entering 'third_party/kleidiai' 2025-09-07T07:36:58.1394562Z Entering 'third_party/mimalloc' 2025-09-07T07:36:58.1427346Z Entering 'third_party/nlohmann' 2025-09-07T07:36:58.1461281Z Entering 'third_party/onnx' 2025-09-07T07:36:58.1506184Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:36:58.1539042Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:36:58.1574745Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:36:58.1609946Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:36:58.1642546Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:36:58.1675609Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:36:58.1707953Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:36:58.1737191Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:36:58.1767987Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:36:58.1800806Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:36:58.1833704Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:36:58.1868234Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:36:58.1917675Z Entering 'third_party/pocketfft' 2025-09-07T07:36:58.1951054Z Entering 'third_party/protobuf' 2025-09-07T07:36:58.1985779Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:36:58.2017189Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:36:58.2050616Z Entering 'third_party/psimd' 2025-09-07T07:36:58.2082303Z Entering 'third_party/pthreadpool' 2025-09-07T07:36:58.2117172Z Entering 'third_party/pybind11' 2025-09-07T07:36:58.2152000Z Entering 'third_party/python-peachpy' 2025-09-07T07:36:58.2185238Z Entering 'third_party/sleef' 2025-09-07T07:36:58.2219021Z Entering 'third_party/tensorpipe' 2025-09-07T07:36:58.2250443Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:36:58.2282675Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:36:58.2313332Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:36:58.2343684Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:36:58.2377018Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:36:58.2427344Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-09-07T07:36:58.2668548Z Entering 'android/libs/fbjni' 2025-09-07T07:36:58.2701146Z Entering 'third_party/FP16' 2025-09-07T07:36:58.2733393Z Entering 'third_party/FXdiv' 2025-09-07T07:36:58.2767309Z Entering 'third_party/NNPACK' 2025-09-07T07:36:58.2799752Z Entering 'third_party/NVTX' 2025-09-07T07:36:58.2833459Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:36:58.2867769Z Entering 'third_party/XNNPACK' 2025-09-07T07:36:58.2912907Z Entering 'third_party/aiter' 2025-09-07T07:36:58.2943830Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:36:58.2982449Z Entering 'third_party/benchmark' 2025-09-07T07:36:58.3012946Z Entering 'third_party/composable_kernel' 2025-09-07T07:36:58.3050001Z Entering 'third_party/cpp-httplib' 2025-09-07T07:36:58.3082416Z Entering 'third_party/cpuinfo' 2025-09-07T07:36:58.3116276Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:36:58.3152503Z Entering 'third_party/cutlass' 2025-09-07T07:36:58.3192472Z Entering 'third_party/fbgemm' 2025-09-07T07:36:58.3227002Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:36:58.3258892Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:36:58.3295632Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:36:58.3327650Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:36:58.3364725Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:36:58.3395736Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:36:58.3424457Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:36:58.3461490Z Entering 'third_party/flash-attention' 2025-09-07T07:36:58.3496326Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:36:58.3530867Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:36:58.3573168Z Entering 'third_party/flatbuffers' 2025-09-07T07:36:58.3608310Z Entering 'third_party/fmt' 2025-09-07T07:36:58.3641358Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:36:58.3677201Z Entering 'third_party/gloo' 2025-09-07T07:36:58.3711427Z Entering 'third_party/googletest' 2025-09-07T07:36:58.3744577Z Entering 'third_party/ideep' 2025-09-07T07:36:58.3777128Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:36:58.3814828Z Entering 'third_party/ittapi' 2025-09-07T07:36:58.3846676Z Entering 'third_party/kineto' 2025-09-07T07:36:58.3880694Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:36:58.3916609Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:36:58.3949942Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:36:58.3980913Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:36:58.4012066Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:36:58.4042022Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:36:58.4077246Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:36:58.4111823Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:36:58.4144903Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:36:58.4178915Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:36:58.4209339Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:36:58.4242881Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:36:58.4279653Z Entering 'third_party/kleidiai' 2025-09-07T07:36:58.4312655Z Entering 'third_party/mimalloc' 2025-09-07T07:36:58.4344369Z Entering 'third_party/nlohmann' 2025-09-07T07:36:58.4378123Z Entering 'third_party/onnx' 2025-09-07T07:36:58.4421711Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:36:58.4454565Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:36:58.4489566Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:36:58.4519434Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:36:58.4551767Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:36:58.4582823Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:36:58.4614836Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:36:58.4649657Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:36:58.4683920Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:36:58.4718627Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:36:58.4753043Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:36:58.4786799Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:36:58.4833762Z Entering 'third_party/pocketfft' 2025-09-07T07:36:58.4866289Z Entering 'third_party/protobuf' 2025-09-07T07:36:58.4899616Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:36:58.4931613Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:36:58.4965098Z Entering 'third_party/psimd' 2025-09-07T07:36:58.4999920Z Entering 'third_party/pthreadpool' 2025-09-07T07:36:58.5032961Z Entering 'third_party/pybind11' 2025-09-07T07:36:58.5064880Z Entering 'third_party/python-peachpy' 2025-09-07T07:36:58.5098297Z Entering 'third_party/sleef' 2025-09-07T07:36:58.5131175Z Entering 'third_party/tensorpipe' 2025-09-07T07:36:58.5164969Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:36:58.5196545Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:36:58.5227407Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:36:58.5260770Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:36:58.5292591Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:36:58.5334297Z ##[endgroup] 2025-09-07T07:36:58.5361199Z [command]/usr/bin/git log -1 --format=%H 2025-09-07T07:36:58.5381359Z 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:36:58.5458800Z ##[group]Run cd "${GITHUB_WORKSPACE}" 2025-09-07T07:36:58.5459045Z cd "${GITHUB_WORKSPACE}" 2025-09-07T07:36:58.5459250Z # Clean stale submodule dirs 2025-09-07T07:36:58.5459454Z if [ -z "${NO_SUDO}" ]; then 2025-09-07T07:36:58.5459688Z  sudo git submodule foreach --recursive git clean -ffdx 2025-09-07T07:36:58.5459930Z else 2025-09-07T07:36:58.5460127Z  git submodule foreach --recursive git clean -ffdx 2025-09-07T07:36:58.5460349Z fi 2025-09-07T07:36:58.5466862Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:36:58.5467098Z env: 2025-09-07T07:36:58.5467253Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:36:58.5467433Z NO_SUDO: true 2025-09-07T07:36:58.5467582Z ##[endgroup] 2025-09-07T07:36:58.5724644Z Entering 'android/libs/fbjni' 2025-09-07T07:36:58.5748083Z Entering 'third_party/FP16' 2025-09-07T07:36:58.5774525Z Entering 'third_party/FXdiv' 2025-09-07T07:36:58.5798354Z Entering 'third_party/NNPACK' 2025-09-07T07:36:58.5824544Z Entering 'third_party/NVTX' 2025-09-07T07:36:58.5856014Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:36:58.5881948Z Entering 'third_party/XNNPACK' 2025-09-07T07:36:58.5969837Z Entering 'third_party/aiter' 2025-09-07T07:36:58.6005958Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:36:58.6090098Z Entering 'third_party/benchmark' 2025-09-07T07:36:58.6114791Z Entering 'third_party/composable_kernel' 2025-09-07T07:36:58.6196086Z Entering 'third_party/cpp-httplib' 2025-09-07T07:36:58.6220350Z Entering 'third_party/cpuinfo' 2025-09-07T07:36:58.6249872Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:36:58.6277329Z Entering 'third_party/cutlass' 2025-09-07T07:36:58.6348265Z Entering 'third_party/fbgemm' 2025-09-07T07:36:58.6389248Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:36:58.6412976Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:36:58.6489160Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:36:58.6513895Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:36:58.6582721Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:36:58.6608460Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:36:58.6632134Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:36:58.6663694Z Entering 'third_party/flash-attention' 2025-09-07T07:36:58.6694243Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:36:58.6760645Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:36:58.6828438Z Entering 'third_party/flatbuffers' 2025-09-07T07:36:58.6879039Z Entering 'third_party/fmt' 2025-09-07T07:36:58.6904457Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:36:58.6928183Z Entering 'third_party/gloo' 2025-09-07T07:36:58.6953053Z Entering 'third_party/googletest' 2025-09-07T07:36:58.6983673Z Entering 'third_party/ideep' 2025-09-07T07:36:58.7008098Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:36:58.7067311Z Entering 'third_party/ittapi' 2025-09-07T07:36:58.7093227Z Entering 'third_party/kineto' 2025-09-07T07:36:58.7119790Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:36:58.7144705Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:36:58.7177037Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:36:58.7201665Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:36:58.7227993Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:36:58.7252811Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:36:58.7279468Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:36:58.7301563Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:36:58.7327890Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:36:58.7358706Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:36:58.7385022Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:36:58.7408150Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:36:58.7435079Z Entering 'third_party/kleidiai' 2025-09-07T07:36:58.7466062Z Entering 'third_party/mimalloc' 2025-09-07T07:36:58.7489712Z Entering 'third_party/nlohmann' 2025-09-07T07:36:58.7521654Z Entering 'third_party/onnx' 2025-09-07T07:36:58.7743442Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:36:58.7768184Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:36:58.7809981Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:36:58.7834337Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:36:58.7861539Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:36:58.7884474Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:36:58.7916893Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:36:58.7942242Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:36:58.7966503Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:36:58.7989896Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:36:58.8023076Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:36:58.8050877Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:36:58.8233063Z Entering 'third_party/pocketfft' 2025-09-07T07:36:58.8256360Z Entering 'third_party/protobuf' 2025-09-07T07:36:58.8309342Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:36:58.8332084Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:36:58.8359724Z Entering 'third_party/psimd' 2025-09-07T07:36:58.8385695Z Entering 'third_party/pthreadpool' 2025-09-07T07:36:58.8410305Z Entering 'third_party/pybind11' 2025-09-07T07:36:58.8437958Z Entering 'third_party/python-peachpy' 2025-09-07T07:36:58.8463629Z Entering 'third_party/sleef' 2025-09-07T07:36:58.8490110Z Entering 'third_party/tensorpipe' 2025-09-07T07:36:58.8518098Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:36:58.8543329Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:36:58.8568096Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:36:58.8592924Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:36:58.8615270Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:36:58.8742677Z Prepare all required actions 2025-09-07T07:36:58.8743158Z Getting action download info 2025-09-07T07:36:59.0278447Z ##[group]Run ./.github/actions/setup-linux 2025-09-07T07:36:59.0278670Z env: 2025-09-07T07:36:59.0278834Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:36:59.0279014Z ##[endgroup] 2025-09-07T07:36:59.0306927Z ##[group]Run set -euo pipefail 2025-09-07T07:36:59.0307204Z set -euo pipefail 2025-09-07T07:36:59.0307421Z function get_ec2_metadata() { 2025-09-07T07:36:59.0307685Z  # Pulled from instance metadata endpoint for EC2 2025-09-07T07:36:59.0308109Z  # see https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/instancedata-data-retrieval.html 2025-09-07T07:36:59.0309248Z  category=$1 2025-09-07T07:36:59.0309512Z  # If it is GCP runner (runner name contains gcp), do not run this 2025-09-07T07:36:59.0309806Z  runner_name_str=i-05ba5cf17983f662f 2025-09-07T07:36:59.0310089Z  if [[ -f /.inarc ]]; then 2025-09-07T07:36:59.0310336Z  echo "ARC Runner, no info on ec2 metadata" 2025-09-07T07:36:59.0310598Z  elif [[ $runner_name_str == *"gcp"* ]]; then 2025-09-07T07:36:59.0310909Z  echo "Runner is from Google Cloud Platform, No info on ec2 metadata" 2025-09-07T07:36:59.0311187Z  else 2025-09-07T07:36:59.0311735Z  curl -H "X-aws-ec2-metadata-token: $(curl -s -X PUT "http://169.254.169.254/latest/api/token" -H "X-aws-ec2-metadata-token-ttl-seconds: 30")" -fsSL "http://169.254.169.254/latest/meta-data/${category}" 2025-09-07T07:36:59.0312281Z  fi 2025-09-07T07:36:59.0312435Z } 2025-09-07T07:36:59.0312636Z echo "ami-id: $(get_ec2_metadata ami-id)" 2025-09-07T07:36:59.0312928Z echo "instance-id: $(get_ec2_metadata instance-id)" 2025-09-07T07:36:59.0313256Z echo "instance-type: $(get_ec2_metadata instance-type)" 2025-09-07T07:36:59.0313518Z echo "system info $(uname -a)" 2025-09-07T07:36:59.0318450Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:36:59.0318699Z env: 2025-09-07T07:36:59.0318860Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:36:59.0319036Z ##[endgroup] 2025-09-07T07:36:59.0425189Z ami-id: ami-05ffe3c48a9991133 2025-09-07T07:36:59.0508117Z instance-id: i-05ba5cf17983f662f 2025-09-07T07:36:59.0586622Z instance-type: c7i.metal-24xl 2025-09-07T07:36:59.0596378Z system info Linux ip-10-0-9-178.ec2.internal 6.1.141-155.222.amzn2023.x86_64 #1 SMP PREEMPT_DYNAMIC Tue Jun 17 10:29:47 UTC 2025 x86_64 x86_64 x86_64 GNU/Linux 2025-09-07T07:36:59.0609731Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:36:59.0610260Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:36:59.0614423Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:36:59.0614669Z env: 2025-09-07T07:36:59.0614822Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:36:59.0614994Z ##[endgroup] 2025-09-07T07:36:59.0678124Z ##[group]Run if systemctl is-active --quiet docker; then 2025-09-07T07:36:59.0678446Z if systemctl is-active --quiet docker; then 2025-09-07T07:36:59.0678729Z  echo "Docker daemon is running..."; 2025-09-07T07:36:59.0678960Z else 2025-09-07T07:36:59.0679212Z  echo "Starting docker daemon..." && sudo systemctl start docker; 2025-09-07T07:36:59.0679498Z fi 2025-09-07T07:36:59.0683334Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:36:59.0683606Z env: 2025-09-07T07:36:59.0683793Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:36:59.0683990Z ##[endgroup] 2025-09-07T07:36:59.0742079Z Docker daemon is running... 2025-09-07T07:36:59.0767542Z ##[group]Run nick-fields/retry@v3.0.0 2025-09-07T07:36:59.0767778Z with: 2025-09-07T07:36:59.0767935Z shell: bash 2025-09-07T07:36:59.0768217Z timeout_minutes: 5 2025-09-07T07:36:59.0768523Z max_attempts: 3 2025-09-07T07:36:59.0768707Z retry_wait_seconds: 30 2025-09-07T07:36:59.0770165Z command: AWS_ACCOUNT_ID=$(aws sts get-caller-identity|grep Account|cut -f4 -d\") aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" # For LF Runners we need to make sure we also login to Meta's ECR docker registry too. META_AWS_ACCOUNT_ID=308535385114 if [ "$AWS_ACCOUNT_ID" != "$META_AWS_ACCOUNT_ID" ] ; then aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$META_AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" fi 2025-09-07T07:36:59.0771622Z polling_interval_seconds: 1 2025-09-07T07:36:59.0771832Z warning_on_retry: true 2025-09-07T07:36:59.0772030Z continue_on_error: false 2025-09-07T07:36:59.0772215Z env: 2025-09-07T07:36:59.0772382Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:36:59.0772554Z AWS_RETRY_MODE: standard 2025-09-07T07:36:59.0772714Z AWS_MAX_ATTEMPTS: 5 2025-09-07T07:36:59.0772873Z AWS_DEFAULT_REGION: us-east-1 2025-09-07T07:36:59.0773054Z ##[endgroup] 2025-09-07T07:36:59.8840602Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-09-07T07:36:59.8840995Z Configure a credential helper to remove this warning. See 2025-09-07T07:36:59.8841351Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-09-07T07:36:59.8841574Z 2025-09-07T07:36:59.8841649Z Login Succeeded 2025-09-07T07:37:00.1385037Z Command completed after 1 attempt(s). 2025-09-07T07:37:00.1426490Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:37:00.1426819Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:37:00.1427092Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:37:00.1433030Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:00.1433260Z env: 2025-09-07T07:37:00.1433421Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:00.1433593Z ##[endgroup] 2025-09-07T07:37:00.1489821Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-09-07T07:37:00.1490166Z # ignore expansion of "docker ps -q" since it could be empty 2025-09-07T07:37:00.1490424Z # shellcheck disable=SC2046 2025-09-07T07:37:00.1490647Z docker stop $(docker ps -q) || true 2025-09-07T07:37:00.1490875Z # Prune all of the docker images 2025-09-07T07:37:00.1491076Z docker system prune -af 2025-09-07T07:37:00.1494853Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:00.1495092Z env: 2025-09-07T07:37:00.1495236Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:00.1495408Z ##[endgroup] 2025-09-07T07:37:00.1942414Z "docker stop" requires at least 1 argument. 2025-09-07T07:37:00.1942761Z See 'docker stop --help'. 2025-09-07T07:37:00.1942999Z 2025-09-07T07:37:00.1943218Z Usage: docker stop [OPTIONS] CONTAINER [CONTAINER...] 2025-09-07T07:37:00.1943401Z 2025-09-07T07:37:00.1943492Z Stop one or more running containers 2025-09-07T07:37:00.2119313Z Total reclaimed space: 0B 2025-09-07T07:37:00.2138138Z ##[group]Run set +e 2025-09-07T07:37:00.2138331Z set +e 2025-09-07T07:37:00.2138494Z set -x 2025-09-07T07:37:00.2138653Z  2025-09-07T07:37:00.2138833Z PT_DOMAIN=download.pytorch.org 2025-09-07T07:37:00.2139189Z # TODO: Flaky access to download.pytorch.org https://github.com/pytorch/pytorch/issues/100400, 2025-09-07T07:37:00.2139629Z # cleaning this up once the issue is fixed. There are more than one resolved IP here, the last 2025-09-07T07:37:00.2139954Z # one is returned at random 2025-09-07T07:37:00.2140214Z RESOLVED_IP=$(dig -4 +short "${PT_DOMAIN}" | tail -n1) 2025-09-07T07:37:00.2140446Z  2025-09-07T07:37:00.2140699Z if [ -z "${RESOLVED_IP}" ]; then 2025-09-07T07:37:00.2140973Z  echo "Couldn't resolve ${PT_DOMAIN}, retrying with Google DNS..." 2025-09-07T07:37:00.2141292Z  RESOLVED_IP=$(dig -4 +short "${PT_DOMAIN}" @8.8.8.8 | tail -n1) 2025-09-07T07:37:00.2141532Z  2025-09-07T07:37:00.2141687Z  if [ -z "${RESOLVED_IP}" ]; then 2025-09-07T07:37:00.2141937Z  echo "Couldn't resolve ${PT_DOMAIN}, exiting..." 2025-09-07T07:37:00.2142173Z  exit 1 2025-09-07T07:37:00.2142335Z  fi 2025-09-07T07:37:00.2142479Z fi 2025-09-07T07:37:00.2142628Z  2025-09-07T07:37:00.2142806Z if grep -r "${PT_DOMAIN}" /etc/hosts; then 2025-09-07T07:37:00.2143121Z  # Clean up any old records first 2025-09-07T07:37:00.2143343Z  sudo sed -i "/${PT_DOMAIN}/d" /etc/hosts 2025-09-07T07:37:00.2143552Z fi 2025-09-07T07:37:00.2143690Z  2025-09-07T07:37:00.2143898Z echo "${RESOLVED_IP} ${PT_DOMAIN}" | sudo tee -a /etc/hosts 2025-09-07T07:37:00.2144134Z cat /etc/hosts 2025-09-07T07:37:00.2148062Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:00.2148301Z env: 2025-09-07T07:37:00.2148458Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:00.2148629Z ##[endgroup] 2025-09-07T07:37:00.2166169Z + PT_DOMAIN=download.pytorch.org 2025-09-07T07:37:00.2172457Z ++ tail -n1 2025-09-07T07:37:00.2172643Z ++ dig -4 +short download.pytorch.org 2025-09-07T07:37:00.2984125Z + RESOLVED_IP=18.160.10.36 2025-09-07T07:37:00.2984510Z + '[' -z 18.160.10.36 ']' 2025-09-07T07:37:00.2984803Z + grep -r download.pytorch.org /etc/hosts 2025-09-07T07:37:00.2994077Z + sudo tee -a /etc/hosts 2025-09-07T07:37:00.2994620Z + echo '18.160.10.36 download.pytorch.org' 2025-09-07T07:37:00.5496367Z 18.160.10.36 download.pytorch.org 2025-09-07T07:37:00.5507220Z + cat /etc/hosts 2025-09-07T07:37:00.5514050Z 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 2025-09-07T07:37:00.5517875Z ::1 localhost6 localhost6.localdomain6 2025-09-07T07:37:00.5518127Z 18.160.10.36 download.pytorch.org 2025-09-07T07:37:00.5610350Z ##[group]Run pytorch/test-infra/.github/actions/calculate-docker-image@main 2025-09-07T07:37:00.5610629Z with: 2025-09-07T07:37:00.5611115Z docker-image-name: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:00.5611641Z use-custom-docker-registry: true 2025-09-07T07:37:00.5611836Z docker-build-dir: .ci/docker 2025-09-07T07:37:00.5612031Z docker-build-script: ./build.sh 2025-09-07T07:37:00.5612218Z working-directory: . 2025-09-07T07:37:00.5612451Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:00.5612681Z force-push: false 2025-09-07T07:37:00.5612840Z env: 2025-09-07T07:37:00.5612984Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:00.5613153Z ##[endgroup] 2025-09-07T07:37:00.5624937Z ##[group]Run set -ex 2025-09-07T07:37:00.5625143Z set -ex 2025-09-07T07:37:00.5625297Z  2025-09-07T07:37:00.5625574Z # If the docker build directory or the build script doesn't exist, the action will 2025-09-07T07:37:00.5625954Z # gracefully return the docker image name as it is. Pulling docker image in Linux 2025-09-07T07:37:00.5626275Z # job could then download the pre-built image as usual 2025-09-07T07:37:00.5626666Z if [[ -d "${DOCKER_BUILD_DIR}" ]] && [[ -f "${DOCKER_BUILD_DIR}/${DOCKER_BUILD_SCRIPT}" ]] && [[ "${USE_CUSTOM_DOCKER_REGISTRY}" == "true" ]]; then 2025-09-07T07:37:00.5627024Z  echo "skip=false" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:00.5627233Z else 2025-09-07T07:37:00.5627407Z  echo "skip=true" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:00.5627666Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:00.5627907Z  2025-09-07T07:37:00.5628228Z  echo "Not using custom ECR registry. Either it was not requested or there is no Docker build script in the ${REPO_NAME} repo..." 2025-09-07T07:37:00.5628574Z  exit 0 2025-09-07T07:37:00.5628717Z fi 2025-09-07T07:37:00.5628858Z  2025-09-07T07:37:00.5629068Z if [[ "${DOCKER_IMAGE_NAME}" == *"${DOCKER_REGISTRY}/${REPO_NAME}"* ]]; then 2025-09-07T07:37:00.5629411Z  # The docker image name already includes the ECR prefix and tag, so we can just 2025-09-07T07:37:00.5629719Z  # use it as it is, but first let's extract the tag 2025-09-07T07:37:00.5629998Z  DOCKER_TAG=$(echo "${DOCKER_IMAGE_NAME}" | awk -F '[:,]' '{print $2}') 2025-09-07T07:37:00.5630376Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:00.5630657Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:00.5630888Z else 2025-09-07T07:37:00.5631062Z  if [[ "${DOCKER_IMAGE_NAME}" == *:* ]]; then 2025-09-07T07:37:00.5631284Z  CUSTOM_TAG_PREFIX=${DOCKER_IMAGE_NAME#*:} 2025-09-07T07:37:00.5631528Z  DOCKER_IMAGE_NAME=${DOCKER_IMAGE_NAME%%:*} 2025-09-07T07:37:00.5631732Z  fi 2025-09-07T07:37:00.5632000Z  DOCKER_TAG=${CUSTOM_TAG_PREFIX:+${CUSTOM_TAG_PREFIX}-}$(git rev-parse HEAD:"${DOCKER_BUILD_DIR}") 2025-09-07T07:37:00.5632339Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:00.5632693Z  echo "docker-image=${DOCKER_REGISTRY}/${REPO_NAME}/${DOCKER_IMAGE_NAME}:${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:00.5633076Z  echo "custom-tag-prefix=${CUSTOM_TAG_PREFIX}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:00.5633324Z fi 2025-09-07T07:37:00.5637514Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:00.5637736Z env: 2025-09-07T07:37:00.5637890Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:00.5638065Z REPO_NAME: pytorch 2025-09-07T07:37:00.5638641Z DOCKER_IMAGE_NAME: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:00.5639151Z DOCKER_BUILD_DIR: .ci/docker 2025-09-07T07:37:00.5639334Z DOCKER_BUILD_SCRIPT: ./build.sh 2025-09-07T07:37:00.5639576Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:00.5639818Z USE_CUSTOM_DOCKER_REGISTRY: true 2025-09-07T07:37:00.5640005Z CUSTOM_TAG_PREFIX: 2025-09-07T07:37:00.5640160Z ##[endgroup] 2025-09-07T07:37:00.5657981Z + [[ -d .ci/docker ]] 2025-09-07T07:37:00.5658211Z + [[ -f .ci/docker/./build.sh ]] 2025-09-07T07:37:00.5658416Z + [[ true == \t\r\u\e ]] 2025-09-07T07:37:00.5658602Z + echo skip=false 2025-09-07T07:37:00.5659324Z + [[ 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 == *\3\0\8\5\3\5\3\8\5\1\1\4\.\d\k\r\.\e\c\r\.\u\s\-\e\a\s\t\-\1\.\a\m\a\z\o\n\a\w\s\.\c\o\m\/\p\y\t\o\r\c\h* ]] 2025-09-07T07:37:00.5665159Z ++ echo 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:00.5665670Z ++ awk -F '[:,]' '{print $2}' 2025-09-07T07:37:00.5681601Z + DOCKER_TAG=pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:00.5682177Z + echo docker-tag=pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:00.5682909Z + echo docker-image=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:00.5698978Z ##[group]Run set +e 2025-09-07T07:37:00.5699190Z set +e 2025-09-07T07:37:00.5699350Z set -x 2025-09-07T07:37:00.5699487Z  2025-09-07T07:37:00.5699636Z login() { 2025-09-07T07:37:00.5699934Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-09-07T07:37:00.5700247Z } 2025-09-07T07:37:00.5700382Z  2025-09-07T07:37:00.5700522Z retry () { 2025-09-07T07:37:00.5700700Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-09-07T07:37:00.5700902Z } 2025-09-07T07:37:00.5701032Z  2025-09-07T07:37:00.5701185Z retry login "${DOCKER_REGISTRY}" 2025-09-07T07:37:00.5701379Z  2025-09-07T07:37:00.5701525Z START_TIME=$(date +%s) 2025-09-07T07:37:00.5701716Z # Wait up to 120 minutes 2025-09-07T07:37:00.5701956Z while [[ $(( $(date +%s) - 7200 )) -lt $START_TIME ]]; do 2025-09-07T07:37:00.5702334Z  # Check if image already exists, if it does then skip building it 2025-09-07T07:37:00.5702635Z  if docker manifest inspect "${DOCKER_IMAGE}"; then 2025-09-07T07:37:00.5702855Z  exit 0 2025-09-07T07:37:00.5703018Z  fi 2025-09-07T07:37:00.5703168Z  2025-09-07T07:37:00.5703416Z  # NB: This flag is used by Docker build workflow to push the image to ECR, so we can 2025-09-07T07:37:00.5703807Z  # use this to differentiate between the Docker build and regular build jobs. For the 2025-09-07T07:37:00.5704188Z  # latter, it will wait for the Docker images to become available before continuing 2025-09-07T07:37:00.5704508Z  if [ "${DOCKER_PUSH:-false}" == "true" ]; then 2025-09-07T07:37:00.5704767Z  # It's a Docker build job, let's build the image 2025-09-07T07:37:00.5704984Z  break 2025-09-07T07:37:00.5705135Z  else 2025-09-07T07:37:00.5705361Z  # It's a regular build job, wait for the image to become available 2025-09-07T07:37:00.5705611Z  sleep 300 2025-09-07T07:37:00.5705786Z  fi 2025-09-07T07:37:00.5705940Z done 2025-09-07T07:37:00.5706091Z  2025-09-07T07:37:00.5706319Z # NB: This part requires a full checkout. Otherwise, the merge base will 2025-09-07T07:37:00.5706726Z # be empty. The default action would be to continue rebuild the image 2025-09-07T07:37:00.5707035Z if [[ "$BASE_REVISION" = "$(git rev-parse HEAD)" ]]; then 2025-09-07T07:37:00.5707309Z  # if we're on the base branch then use the parent commit 2025-09-07T07:37:00.5707552Z  MERGE_BASE=$(git rev-parse HEAD~) 2025-09-07T07:37:00.5707746Z else 2025-09-07T07:37:00.5707943Z  # otherwise we're on a PR, so use the most recent base commit 2025-09-07T07:37:00.5708222Z  MERGE_BASE=$(git merge-base HEAD "$BASE_REVISION") 2025-09-07T07:37:00.5708441Z fi 2025-09-07T07:37:00.5708588Z  2025-09-07T07:37:00.5708738Z if [[ -z "${MERGE_BASE}" ]]; then 2025-09-07T07:37:00.5708963Z  echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:00.5709168Z  2025-09-07T07:37:00.5709443Z  echo "Finding merge base only works with full checkout, please set fetch-depth to 0, continuing ..." 2025-09-07T07:37:00.5709748Z  exit 0 2025-09-07T07:37:00.5709897Z fi 2025-09-07T07:37:00.5710035Z  2025-09-07T07:37:00.5710227Z if ! git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}"; then 2025-09-07T07:37:00.5710615Z  echo "Directory '${DOCKER_BUILD_DIR}' not found in commit $MERGE_BASE, you should rebase onto a more recent commit" 2025-09-07T07:37:00.5710932Z  exit 1 2025-09-07T07:37:00.5711078Z fi 2025-09-07T07:37:00.5711214Z  2025-09-07T07:37:00.5711437Z PREVIOUS_DOCKER_TAG=$(git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}") 2025-09-07T07:37:00.5711800Z # If no image exists but the hash is the same as the previous hash then we should error out here 2025-09-07T07:37:00.5712137Z if [[ "${PREVIOUS_DOCKER_TAG}" == "${DOCKER_TAG}" ]]; then 2025-09-07T07:37:00.5712522Z  echo "WARNING: Something has gone wrong and the previous image isn't available for the merge-base of your branch" 2025-09-07T07:37:00.5712948Z  echo " Will re-build docker image to store in local cache, TTS may be longer" 2025-09-07T07:37:00.5713208Z fi 2025-09-07T07:37:00.5713336Z  2025-09-07T07:37:00.5713504Z echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:00.5717174Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:00.5717413Z env: 2025-09-07T07:37:00.5717574Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:00.5717763Z DOCKER_BUILD_DIR: .ci/docker 2025-09-07T07:37:00.5717984Z BASE_REVISION: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:37:00.5718593Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:00.5719280Z DOCKER_TAG: pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:00.5719712Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:00.5719958Z DOCKER_PUSH: 2025-09-07T07:37:00.5720108Z ##[endgroup] 2025-09-07T07:37:00.5737420Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:00.5737684Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:00.5740845Z + aws ecr get-login-password --region us-east-1 2025-09-07T07:37:00.5741446Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:00.9304107Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-09-07T07:37:00.9304507Z Configure a credential helper to remove this warning. See 2025-09-07T07:37:00.9304977Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-09-07T07:37:00.9305204Z 2025-09-07T07:37:00.9305272Z Login Succeeded 2025-09-07T07:37:00.9316605Z ++ date +%s 2025-09-07T07:37:00.9323049Z + START_TIME=1757230620 2025-09-07T07:37:00.9325485Z ++ date +%s 2025-09-07T07:37:00.9332260Z + [[ 1757223420 -lt 1757230620 ]] 2025-09-07T07:37:00.9332835Z + docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:01.1390958Z { 2025-09-07T07:37:01.1391215Z "schemaVersion": 2, 2025-09-07T07:37:01.1391510Z "mediaType": "application/vnd.docker.distribution.manifest.v2+json", 2025-09-07T07:37:01.1391796Z "config": { 2025-09-07T07:37:01.1392028Z "mediaType": "application/vnd.docker.container.image.v1+json", 2025-09-07T07:37:01.1392301Z "size": 30269, 2025-09-07T07:37:01.1392567Z "digest": "sha256:662d8c9dfc7db2f5d004293de4f2b7647941dee4c916479ef082d17fcdfd9c47" 2025-09-07T07:37:01.1392852Z }, 2025-09-07T07:37:01.1392990Z "layers": [ 2025-09-07T07:37:01.1393132Z { 2025-09-07T07:37:01.1393337Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1393592Z "size": 30448359, 2025-09-07T07:37:01.1393884Z "digest": "sha256:e6fdc8487bfe6d764301ef3634bc6c043841dc3ab05ca14f81e69c0f92562d46" 2025-09-07T07:37:01.1394179Z }, 2025-09-07T07:37:01.1394304Z { 2025-09-07T07:37:01.1394513Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1394756Z "size": 1554, 2025-09-07T07:37:01.1395016Z "digest": "sha256:18a5ee5b0e2e283bf6d7b9c4c312b0448c75eff1c43446c22c5139a3aeec97fe" 2025-09-07T07:37:01.1395291Z }, 2025-09-07T07:37:01.1395412Z { 2025-09-07T07:37:01.1395616Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1395871Z "size": 313297813, 2025-09-07T07:37:01.1396135Z "digest": "sha256:572424b92528ee46c84fdf3e9e1f5fd75e302621ad75dcf4257ad06778885094" 2025-09-07T07:37:01.1396399Z }, 2025-09-07T07:37:01.1396525Z { 2025-09-07T07:37:01.1396724Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1396966Z "size": 793, 2025-09-07T07:37:01.1397223Z "digest": "sha256:1c35b7d4b67c6769f59f96a643d69c214c5b00291a4968cdd395eedbce82b9c0" 2025-09-07T07:37:01.1397494Z }, 2025-09-07T07:37:01.1397622Z { 2025-09-07T07:37:01.1397820Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1398052Z "size": 106, 2025-09-07T07:37:01.1398313Z "digest": "sha256:68c20f3c23bb0bddb9b69e6ce2e45bcd5b1fcfd9b37dbe3de26b8a5f0e81ff13" 2025-09-07T07:37:01.1398596Z }, 2025-09-07T07:37:01.1398720Z { 2025-09-07T07:37:01.1398912Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1399153Z "size": 704, 2025-09-07T07:37:01.1399618Z "digest": "sha256:7efa39950d3273a15b20bc5f6659373b2b4eb62e36328d96b289834c48d2e408" 2025-09-07T07:37:01.1399888Z }, 2025-09-07T07:37:01.1400010Z { 2025-09-07T07:37:01.1400212Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1400453Z "size": 1214, 2025-09-07T07:37:01.1400703Z "digest": "sha256:a10eb16a7271e996ea9f1d769ba6bd2ec69358f2a79cf26649595a8cea38275f" 2025-09-07T07:37:01.1400972Z }, 2025-09-07T07:37:01.1401101Z { 2025-09-07T07:37:01.1401472Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1401778Z "size": 485, 2025-09-07T07:37:01.1402015Z "digest": "sha256:7d52cf57965449440c17f257fe4c522f9685019961eaa9853d7c820cfe39f5cc" 2025-09-07T07:37:01.1402284Z }, 2025-09-07T07:37:01.1402416Z { 2025-09-07T07:37:01.1402615Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1402848Z "size": 110343705, 2025-09-07T07:37:01.1403107Z "digest": "sha256:cb6a20fcf4e24ec2e1f72ecf361b26e058f3e6194947a9b3a25312223d43516e" 2025-09-07T07:37:01.1403383Z }, 2025-09-07T07:37:01.1403511Z { 2025-09-07T07:37:01.1403699Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1403938Z "size": 4787, 2025-09-07T07:37:01.1404184Z "digest": "sha256:46fb6a8b3e1d4eac9b3a21577824410003ed38f194b4b1486b747e324b32ef6a" 2025-09-07T07:37:01.1404455Z }, 2025-09-07T07:37:01.1404664Z { 2025-09-07T07:37:01.1404872Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1405112Z "size": 1709, 2025-09-07T07:37:01.1405369Z "digest": "sha256:5ad6977cc38e4ea8a6545d6a4fc0e2fdde705a7af96eb496cfe20f264fbc1e74" 2025-09-07T07:37:01.1405644Z }, 2025-09-07T07:37:01.1405772Z { 2025-09-07T07:37:01.1405973Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1406215Z "size": 724, 2025-09-07T07:37:01.1406455Z "digest": "sha256:da63046995a2e510b7146776371a14bff4b31002cc3ef0322e45a3932fba2031" 2025-09-07T07:37:01.1406725Z }, 2025-09-07T07:37:01.1406853Z { 2025-09-07T07:37:01.1407045Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1407277Z "size": 543, 2025-09-07T07:37:01.1407523Z "digest": "sha256:78243fdb9906cb588921ddaa67a3ca915aa9447ca675faac1a9ebc420a561d83" 2025-09-07T07:37:01.1407793Z }, 2025-09-07T07:37:01.1407914Z { 2025-09-07T07:37:01.1408108Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1408351Z "size": 3395447162, 2025-09-07T07:37:01.1408613Z "digest": "sha256:6f70d5d50abaab8988f460b5590d92b6d1d340575ddee981662c24034d7d20af" 2025-09-07T07:37:01.1408886Z }, 2025-09-07T07:37:01.1409001Z { 2025-09-07T07:37:01.1409202Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1409517Z + exit 0 2025-09-07T07:37:01.1409659Z "size": 32, 2025-09-07T07:37:01.1409911Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:01.1410182Z }, 2025-09-07T07:37:01.1410331Z { 2025-09-07T07:37:01.1410534Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1410777Z "size": 380, 2025-09-07T07:37:01.1411030Z "digest": "sha256:69715d3ad3c493436abde51f5a575e79f7d55b46c653f5607f3c7722ad9a05db" 2025-09-07T07:37:01.1411294Z }, 2025-09-07T07:37:01.1411426Z { 2025-09-07T07:37:01.1411622Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1411868Z "size": 235844, 2025-09-07T07:37:01.1412115Z "digest": "sha256:7ace90c063f3f3ce8f04b541afe935088868930e5c074824af2b2c327779a3b5" 2025-09-07T07:37:01.1412389Z }, 2025-09-07T07:37:01.1412514Z { 2025-09-07T07:37:01.1412708Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1412940Z "size": 230, 2025-09-07T07:37:01.1413189Z "digest": "sha256:acbd5447dd1406dab8e46234f6a034a75ad9794f76c24f817b0ecf28b6a69c78" 2025-09-07T07:37:01.1414057Z }, 2025-09-07T07:37:01.1414195Z { 2025-09-07T07:37:01.1414391Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1414644Z "size": 3396092, 2025-09-07T07:37:01.1414906Z "digest": "sha256:744523d9b7f5a3e7abfc646c2d5222e7379024242430b93cb4b8093574e69022" 2025-09-07T07:37:01.1415180Z }, 2025-09-07T07:37:01.1415306Z { 2025-09-07T07:37:01.1415518Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1415774Z "size": 1477, 2025-09-07T07:37:01.1416036Z "digest": "sha256:5bd615a7b945084e11bcb40190f9d6e50367297237146df7b008fa8c668f29c8" 2025-09-07T07:37:01.1416308Z }, 2025-09-07T07:37:01.1416440Z { 2025-09-07T07:37:01.1416644Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1416889Z "size": 482, 2025-09-07T07:37:01.1417148Z "digest": "sha256:f4986a00e3aecf1d56beaada7aba8c49fbb3683db3c99790ab0aa4caaa34f76f" 2025-09-07T07:37:01.1417434Z }, 2025-09-07T07:37:01.1417566Z { 2025-09-07T07:37:01.1417770Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1418009Z "size": 196, 2025-09-07T07:37:01.1418264Z "digest": "sha256:21902f6e4f8cb76c82e755b8fc9f72e1912bf925ab345ab5b4cc2210f4887a64" 2025-09-07T07:37:01.1418537Z }, 2025-09-07T07:37:01.1418669Z { 2025-09-07T07:37:01.1418865Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1419161Z "size": 608, 2025-09-07T07:37:01.1419417Z "digest": "sha256:d80602abf3ccf0c0b527848a403dfde36e1cf1db1416852385feda5c44bf4363" 2025-09-07T07:37:01.1419696Z }, 2025-09-07T07:37:01.1419821Z { 2025-09-07T07:37:01.1420029Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1420275Z "size": 226, 2025-09-07T07:37:01.1420528Z "digest": "sha256:3c51bf0bc362d34a17911f73c5146cbd668c4d1cf1b944cbf40a604d71cd623a" 2025-09-07T07:37:01.1420806Z }, 2025-09-07T07:37:01.1420932Z { 2025-09-07T07:37:01.1421137Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1421389Z "size": 828, 2025-09-07T07:37:01.1421648Z "digest": "sha256:119ab3bceafa6f2cab4b1f71161195139792990263ee8de82230c6284f0ae20a" 2025-09-07T07:37:01.1421926Z }, 2025-09-07T07:37:01.1422059Z { 2025-09-07T07:37:01.1422263Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1422507Z "size": 32, 2025-09-07T07:37:01.1422756Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:01.1423037Z }, 2025-09-07T07:37:01.1423169Z { 2025-09-07T07:37:01.1423367Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1423610Z "size": 104, 2025-09-07T07:37:01.1423865Z "digest": "sha256:af8eadc9eaabdaf6c5e01031d63061605327153e07568ddd159966ecea75cd07" 2025-09-07T07:37:01.1424149Z }, 2025-09-07T07:37:01.1424278Z { 2025-09-07T07:37:01.1424476Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1424731Z "size": 1495, 2025-09-07T07:37:01.1424980Z "digest": "sha256:e7769b0d7a8262f3cc32a9d96080de5318dac3d2617e10508a167e689016e40c" 2025-09-07T07:37:01.1425253Z }, 2025-09-07T07:37:01.1425375Z { 2025-09-07T07:37:01.1425584Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1425832Z "size": 453908015, 2025-09-07T07:37:01.1426096Z "digest": "sha256:ba263639b0f4634277ef3b8903e3457ac27ce012f1bbeeeeb773191c2c3b222b" 2025-09-07T07:37:01.1426361Z }, 2025-09-07T07:37:01.1426493Z { 2025-09-07T07:37:01.1426696Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1426939Z "size": 164, 2025-09-07T07:37:01.1427183Z "digest": "sha256:a5ab7a280382a797dd5ba6a6716f667a231540ad1e0e7c8ba48bb24d5ab80ef0" 2025-09-07T07:37:01.1427458Z }, 2025-09-07T07:37:01.1427590Z { 2025-09-07T07:37:01.1427791Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1428059Z "size": 346, 2025-09-07T07:37:01.1428310Z "digest": "sha256:80b2232d952f55c3662cffd657ba30fe825f08dfcc5bbea13e2bc6de4482b7e4" 2025-09-07T07:37:01.1428582Z }, 2025-09-07T07:37:01.1428706Z { 2025-09-07T07:37:01.1428897Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1429138Z "size": 32, 2025-09-07T07:37:01.1429395Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:01.1429666Z }, 2025-09-07T07:37:01.1429784Z { 2025-09-07T07:37:01.1429980Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1430221Z "size": 106, 2025-09-07T07:37:01.1430466Z "digest": "sha256:cc93cd65e90f0a9c50194579c93e96897f4e582b9777a1c4d7df7b913ddcdded" 2025-09-07T07:37:01.1430728Z }, 2025-09-07T07:37:01.1430856Z { 2025-09-07T07:37:01.1431059Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1431301Z "size": 425, 2025-09-07T07:37:01.1431545Z "digest": "sha256:0eed4c15712bc470dac7df87e33b3570a1510344019dd9cc0e95b8beb1f98372" 2025-09-07T07:37:01.1431816Z }, 2025-09-07T07:37:01.1431946Z { 2025-09-07T07:37:01.1432143Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1432382Z "size": 19309387, 2025-09-07T07:37:01.1432632Z "digest": "sha256:092516f71fe325518f9737f105bcd65c40cd35c3019098889757e2c84c03c8a8" 2025-09-07T07:37:01.1432932Z }, 2025-09-07T07:37:01.1433057Z { 2025-09-07T07:37:01.1433248Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1433490Z "size": 108, 2025-09-07T07:37:01.1433736Z "digest": "sha256:8c0825014a6270f765ff514da8583d55874f3278bef76e5617e29115f91ee654" 2025-09-07T07:37:01.1434005Z }, 2025-09-07T07:37:01.1434122Z { 2025-09-07T07:37:01.1434321Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1434570Z "size": 636, 2025-09-07T07:37:01.1434822Z "digest": "sha256:8e0d2f63da0a8ff07657d7e06cdbc1ad9d5db95614d640a9f7a9aa8c30c9986d" 2025-09-07T07:37:01.1435099Z }, 2025-09-07T07:37:01.1435222Z { 2025-09-07T07:37:01.1435422Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1435664Z "size": 724, 2025-09-07T07:37:01.1435902Z "digest": "sha256:da63046995a2e510b7146776371a14bff4b31002cc3ef0322e45a3932fba2031" 2025-09-07T07:37:01.1436170Z }, 2025-09-07T07:37:01.1436292Z { 2025-09-07T07:37:01.1436491Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1436726Z "size": 148, 2025-09-07T07:37:01.1436973Z "digest": "sha256:73aae7958ba1a16c5f5625d39b06208e1def8c7816bb75028bf0845f553a5068" 2025-09-07T07:37:01.1437245Z }, 2025-09-07T07:37:01.1437369Z { 2025-09-07T07:37:01.1437560Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1437801Z "size": 136, 2025-09-07T07:37:01.1438045Z "digest": "sha256:ac6077ec9fa50fc0822d387d2ee35e1b6f1f56612402fe7195378180b25087bc" 2025-09-07T07:37:01.1438321Z }, 2025-09-07T07:37:01.1438437Z { 2025-09-07T07:37:01.1438635Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1438879Z "size": 140, 2025-09-07T07:37:01.1439132Z "digest": "sha256:bf4ee4e45e92ef179f7fc64e2c7c6755905a969c37cf82c39aafbadd9290ff04" 2025-09-07T07:37:01.1439399Z }, 2025-09-07T07:37:01.1439525Z { 2025-09-07T07:37:01.1439725Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1439975Z "size": 18617175577, 2025-09-07T07:37:01.1440240Z "digest": "sha256:c1b766f9b961bcc863d6f89d623815fd7dfe9797ddcfd5d15ef06ffe7d177359" 2025-09-07T07:37:01.1440515Z }, 2025-09-07T07:37:01.1440640Z { 2025-09-07T07:37:01.1440840Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1441077Z "size": 223, 2025-09-07T07:37:01.1441334Z "digest": "sha256:6e726ef07b5d5cfe2fb9f06d43fc931fc64c381fd37eaf0c169e0dd84796f152" 2025-09-07T07:37:01.1441613Z }, 2025-09-07T07:37:01.1441779Z { 2025-09-07T07:37:01.1441970Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1442213Z "size": 274477524, 2025-09-07T07:37:01.1442472Z "digest": "sha256:364070434a64fa913f3907ada910a4051707e693e0e6124f57bc97aa57791da1" 2025-09-07T07:37:01.1442735Z }, 2025-09-07T07:37:01.1442854Z { 2025-09-07T07:37:01.1443052Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1443296Z "size": 6451569004, 2025-09-07T07:37:01.1443555Z "digest": "sha256:71f708151a84685fc366b85e914dac9f5279313eff07358d79ecaaeecb0f1c42" 2025-09-07T07:37:01.1443817Z }, 2025-09-07T07:37:01.1443944Z { 2025-09-07T07:37:01.1444146Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1444385Z "size": 129, 2025-09-07T07:37:01.1444630Z "digest": "sha256:622d8cfb39ea4dda608d2819c6a9de45df81b6f8319ee8ab4a24c36d81b9a132" 2025-09-07T07:37:01.1444908Z }, 2025-09-07T07:37:01.1445039Z { 2025-09-07T07:37:01.1445243Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1445473Z "size": 778, 2025-09-07T07:37:01.1445719Z "digest": "sha256:284119a92cb13dacff06926444aab4f99756039acb48abba7b75d35c367ed3f1" 2025-09-07T07:37:01.1445984Z }, 2025-09-07T07:37:01.1446109Z { 2025-09-07T07:37:01.1446301Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1446575Z "size": 724, 2025-09-07T07:37:01.1446818Z "digest": "sha256:da63046995a2e510b7146776371a14bff4b31002cc3ef0322e45a3932fba2031" 2025-09-07T07:37:01.1447083Z }, 2025-09-07T07:37:01.1447202Z { 2025-09-07T07:37:01.1447401Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1447640Z "size": 140, 2025-09-07T07:37:01.1447880Z "digest": "sha256:96695940d842555623cfe4fb7b52e949423e8c8f383e55d02363e7e5c5804afa" 2025-09-07T07:37:01.1448135Z }, 2025-09-07T07:37:01.1448258Z { 2025-09-07T07:37:01.1448456Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1448696Z "size": 32, 2025-09-07T07:37:01.1448941Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:01.1449216Z }, 2025-09-07T07:37:01.1449343Z { 2025-09-07T07:37:01.1449539Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1449771Z "size": 160, 2025-09-07T07:37:01.1450023Z "digest": "sha256:7ddca6c4c050460204097ba875dc0fa03eca6265122a18c0b8dc5504152aea53" 2025-09-07T07:37:01.1450295Z }, 2025-09-07T07:37:01.1450417Z { 2025-09-07T07:37:01.1450608Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1450850Z "size": 1012, 2025-09-07T07:37:01.1451115Z "digest": "sha256:a95e1f2f1aadef03514a7cdbdac1fe83d4eebedbb80df9be868a223f27e1c263" 2025-09-07T07:37:01.1451401Z }, 2025-09-07T07:37:01.1451530Z { 2025-09-07T07:37:01.1451729Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1451971Z "size": 724, 2025-09-07T07:37:01.1452215Z "digest": "sha256:da63046995a2e510b7146776371a14bff4b31002cc3ef0322e45a3932fba2031" 2025-09-07T07:37:01.1452471Z }, 2025-09-07T07:37:01.1452601Z { 2025-09-07T07:37:01.1452797Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1453037Z "size": 135, 2025-09-07T07:37:01.1453274Z "digest": "sha256:8085756b0cc0f9588f23a73c27840a5dff48cc18c3a2f0311e4d1ef291855679" 2025-09-07T07:37:01.1453549Z }, 2025-09-07T07:37:01.1453677Z { 2025-09-07T07:37:01.1453875Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1454109Z "size": 32, 2025-09-07T07:37:01.1454359Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:01.1454630Z }, 2025-09-07T07:37:01.1454754Z { 2025-09-07T07:37:01.1454944Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1455189Z "size": 158, 2025-09-07T07:37:01.1455496Z "digest": "sha256:7e9ff0c6f103b18756f01c60b4d57a951660f17bffb1810b330e3ff703caf216" 2025-09-07T07:37:01.1455775Z }, 2025-09-07T07:37:01.1455899Z { 2025-09-07T07:37:01.1456102Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1456350Z "size": 1369, 2025-09-07T07:37:01.1456611Z "digest": "sha256:a625cbbc05b983aeb4c28702a4a5b65c68191ab1b8d17978f7d98cc17ddf3c52" 2025-09-07T07:37:01.1456887Z }, 2025-09-07T07:37:01.1457020Z { 2025-09-07T07:37:01.1457223Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1457472Z "size": 32, 2025-09-07T07:37:01.1457718Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:01.1457996Z }, 2025-09-07T07:37:01.1458129Z { 2025-09-07T07:37:01.1458332Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1458580Z "size": 136, 2025-09-07T07:37:01.1458839Z "digest": "sha256:4e28486424310870c8d6815524440f17c6e0afe7572eaa173a811b98b4920bed" 2025-09-07T07:37:01.1459117Z }, 2025-09-07T07:37:01.1459251Z { 2025-09-07T07:37:01.1459455Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1459711Z "size": 380, 2025-09-07T07:37:01.1459976Z "digest": "sha256:5e944f1ed1bef9442f5b1b86225d3958ea8f2f7f4c6aa7b92dc5d0c810c260bc" 2025-09-07T07:37:01.1460264Z }, 2025-09-07T07:37:01.1460428Z { 2025-09-07T07:37:01.1460633Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1460881Z "size": 32, 2025-09-07T07:37:01.1461134Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:01.1461413Z }, 2025-09-07T07:37:01.1461538Z { 2025-09-07T07:37:01.1461741Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1461993Z "size": 104, 2025-09-07T07:37:01.1462243Z "digest": "sha256:41619248f604c60e038a02bfd462af96ee2996b77be5f59f05e9ac5fe4790e5a" 2025-09-07T07:37:01.1462518Z }, 2025-09-07T07:37:01.1462650Z { 2025-09-07T07:37:01.1462851Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1463095Z "size": 407, 2025-09-07T07:37:01.1486053Z "digest": "sha256:be86f8c4f654b9ae64a20eb7f960e6ce4baa5b46e0a1f5e1312b11492a40bcd4" 2025-09-07T07:37:01.1486371Z }, 2025-09-07T07:37:01.1486508Z { 2025-09-07T07:37:01.1486746Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1487014Z "size": 32, 2025-09-07T07:37:01.1487292Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:01.1487566Z }, 2025-09-07T07:37:01.1487701Z { 2025-09-07T07:37:01.1487914Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1488164Z "size": 109, 2025-09-07T07:37:01.1488426Z "digest": "sha256:ef1340e22a4bc8cf42e1d40961cb32d183cd3da8f0b785b5425c32ee067690c1" 2025-09-07T07:37:01.1488719Z }, 2025-09-07T07:37:01.1488851Z { 2025-09-07T07:37:01.1489064Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1489305Z "size": 1897, 2025-09-07T07:37:01.1489565Z "digest": "sha256:da8d8b696333cbf6b9f339ab859639c905d6752d7e65fea14c23c3c2dcba553e" 2025-09-07T07:37:01.1489841Z }, 2025-09-07T07:37:01.1489971Z { 2025-09-07T07:37:01.1490167Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1490412Z "size": 243443118, 2025-09-07T07:37:01.1490680Z "digest": "sha256:386b0c49c4982a821fb6f427fbc7d9c7d2012e97c96a514a9c7a09304e76b935" 2025-09-07T07:37:01.1490957Z }, 2025-09-07T07:37:01.1491076Z { 2025-09-07T07:37:01.1491277Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1491524Z "size": 106, 2025-09-07T07:37:01.1491789Z "digest": "sha256:2b1d0ea7efe0bf86e86df804d2cddbf83b113fdecd03f3ddfca728da30546f34" 2025-09-07T07:37:01.1492062Z }, 2025-09-07T07:37:01.1492317Z { 2025-09-07T07:37:01.1492519Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1492767Z "size": 163, 2025-09-07T07:37:01.1493012Z "digest": "sha256:04c04be7408f20625b1bd8454e5a08c91fcf04d4f79ab3ec1b75ae6b1824174d" 2025-09-07T07:37:01.1493289Z }, 2025-09-07T07:37:01.1493421Z { 2025-09-07T07:37:01.1493621Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1493863Z "size": 7943, 2025-09-07T07:37:01.1494120Z "digest": "sha256:f8690caa3ac5e845f2dcc25ad12815b5c7452285c3838a87c780bd03ecf072a3" 2025-09-07T07:37:01.1494391Z }, 2025-09-07T07:37:01.1494517Z { 2025-09-07T07:37:01.1494705Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1494949Z "size": 8074, 2025-09-07T07:37:01.1495196Z "digest": "sha256:2908d6baaa6b21331dee5f210472cae0874d22b98b0a35420cad4fd753ed215f" 2025-09-07T07:37:01.1495469Z }, 2025-09-07T07:37:01.1495585Z { 2025-09-07T07:37:01.1495785Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1496036Z "size": 303, 2025-09-07T07:37:01.1496278Z "digest": "sha256:37e2336101eba2c73995d34431e4fae8782d9e9700c42621777922490b2158ed" 2025-09-07T07:37:01.1496534Z }, 2025-09-07T07:37:01.1496659Z { 2025-09-07T07:37:01.1496858Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1497098Z "size": 32, 2025-09-07T07:37:01.1497400Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:01.1497681Z }, 2025-09-07T07:37:01.1497811Z { 2025-09-07T07:37:01.1498010Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1498243Z "size": 108, 2025-09-07T07:37:01.1498487Z "digest": "sha256:f1ac881fde33994861be4324231269058643168b9aee60c699552d0d92d965da" 2025-09-07T07:37:01.1498752Z }, 2025-09-07T07:37:01.1498881Z { 2025-09-07T07:37:01.1499071Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1499325Z "size": 54145699, 2025-09-07T07:37:01.1499586Z "digest": "sha256:43b14c67347e2813c5f63e928c14db60dbb35c330ccc865510cf79739d8b78a1" 2025-09-07T07:37:01.1499859Z }, 2025-09-07T07:37:01.1499980Z { 2025-09-07T07:37:01.1500184Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:01.1500428Z "size": 32, 2025-09-07T07:37:01.1500678Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:01.1500941Z } 2025-09-07T07:37:01.1501069Z ] 2025-09-07T07:37:01.1501198Z } 2025-09-07T07:37:01.1518395Z ##[group]Run set -eux 2025-09-07T07:37:01.1518586Z set -eux 2025-09-07T07:37:01.1518847Z # It's ok if this steps fails, it would then be an anonymous user like what we used to have 2025-09-07T07:37:01.1519487Z aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token | jq --raw-output '.SecretString' | jq -r .docker_hub_readonly_token | docker login --username pytorchbot --password-stdin || true 2025-09-07T07:37:01.1524464Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:01.1524697Z env: 2025-09-07T07:37:01.1524857Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:01.1525045Z ##[endgroup] 2025-09-07T07:37:01.1548121Z + aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token 2025-09-07T07:37:01.1548451Z + jq --raw-output .SecretString 2025-09-07T07:37:01.1548716Z + docker login --username pytorchbot --password-stdin 2025-09-07T07:37:01.1548970Z + jq -r .docker_hub_readonly_token 2025-09-07T07:37:01.5743126Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-09-07T07:37:01.5743506Z Configure a credential helper to remove this warning. See 2025-09-07T07:37:01.5743747Z Login Succeeded 2025-09-07T07:37:01.5744032Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-09-07T07:37:01.5744250Z 2025-09-07T07:37:01.5799705Z ##[group]Run tag=${ECR_DOCKER_IMAGE##*:} 2025-09-07T07:37:01.5801970Z tag=${ECR_DOCKER_IMAGE##*:} 2025-09-07T07:37:01.5802224Z echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}" 2025-09-07T07:37:01.5806482Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:01.5806723Z env: 2025-09-07T07:37:01.5806878Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:01.5807409Z ECR_DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:01.5807924Z ##[endgroup] 2025-09-07T07:37:01.5826754Z docker pull ghcr.io/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:01.5855414Z ##[group]Run pytorch/test-infra/.github/actions/pull-docker-image@main 2025-09-07T07:37:01.5855695Z with: 2025-09-07T07:37:01.5856245Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:01.5856878Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:01.5857142Z env: 2025-09-07T07:37:01.5857307Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:01.5857491Z ##[endgroup] 2025-09-07T07:37:01.5889215Z ##[group]Run set -x 2025-09-07T07:37:01.5889397Z set -x 2025-09-07T07:37:01.5889549Z set +e 2025-09-07T07:37:01.5889697Z  2025-09-07T07:37:01.5889833Z login() { 2025-09-07T07:37:01.5890124Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-09-07T07:37:01.5890425Z } 2025-09-07T07:37:01.5890571Z  2025-09-07T07:37:01.5890740Z retry () { 2025-09-07T07:37:01.5890918Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-09-07T07:37:01.5891107Z } 2025-09-07T07:37:01.5891245Z  2025-09-07T07:37:01.5891398Z retry login "${DOCKER_REGISTRY}" 2025-09-07T07:37:01.5891591Z  2025-09-07T07:37:01.5891869Z IMAGE_SIZE=$(docker manifest inspect "${DOCKER_IMAGE}" | jq '[.layers[].size, .config.size] | add / 1024 / 1024') 2025-09-07T07:37:01.5892239Z echo "Compressed size of image in MB: ${IMAGE_SIZE}" 2025-09-07T07:37:01.5892461Z  2025-09-07T07:37:01.5892599Z set -e 2025-09-07T07:37:01.5892805Z # ignore output since only exit code is used for conditional 2025-09-07T07:37:01.5893096Z # only pull docker image if it's not available locally 2025-09-07T07:37:01.5893407Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2025-09-07T07:37:01.5893701Z  retry docker pull "${DOCKER_IMAGE}" 2025-09-07T07:37:01.5893899Z fi 2025-09-07T07:37:01.5897728Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:01.5897948Z env: 2025-09-07T07:37:01.5898094Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:01.5898599Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:01.5899146Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:01.5899375Z ##[endgroup] 2025-09-07T07:37:01.5915970Z + set +e 2025-09-07T07:37:01.5916361Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:01.5916640Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:01.5920471Z + aws ecr get-login-password --region us-east-1 2025-09-07T07:37:01.5920795Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:01.9365265Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-09-07T07:37:01.9365637Z Configure a credential helper to remove this warning. See 2025-09-07T07:37:01.9365982Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-09-07T07:37:01.9366385Z 2025-09-07T07:37:01.9370585Z Login Succeeded 2025-09-07T07:37:01.9381427Z ++ jq '[.layers[].size, .config.size] | add / 1024 / 1024' 2025-09-07T07:37:01.9382289Z ++ docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:02.1445609Z + IMAGE_SIZE=28579.020259857178 2025-09-07T07:37:02.1445859Z + echo 'Compressed size of image in MB: 28579.020259857178' 2025-09-07T07:37:02.1446105Z + set -e 2025-09-07T07:37:02.1446778Z + docker inspect --type=image 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:02.1447356Z Compressed size of image in MB: 28579.020259857178 2025-09-07T07:37:02.1558620Z + retry docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:02.1559487Z + docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:02.3863333Z pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77: Pulling from pytorch/ci-image 2025-09-07T07:37:02.3863914Z e6fdc8487bfe: Pulling fs layer 2025-09-07T07:37:02.3864152Z 18a5ee5b0e2e: Pulling fs layer 2025-09-07T07:37:02.3864362Z 572424b92528: Pulling fs layer 2025-09-07T07:37:02.3864548Z 1c35b7d4b67c: Pulling fs layer 2025-09-07T07:37:02.3864721Z 68c20f3c23bb: Pulling fs layer 2025-09-07T07:37:02.3864901Z 7efa39950d32: Pulling fs layer 2025-09-07T07:37:02.3865092Z a10eb16a7271: Pulling fs layer 2025-09-07T07:37:02.3865273Z 7d52cf579654: Pulling fs layer 2025-09-07T07:37:02.3865442Z cb6a20fcf4e2: Pulling fs layer 2025-09-07T07:37:02.3865680Z 46fb6a8b3e1d: Pulling fs layer 2025-09-07T07:37:02.3865913Z 5ad6977cc38e: Pulling fs layer 2025-09-07T07:37:02.3866133Z da63046995a2: Pulling fs layer 2025-09-07T07:37:02.3866317Z 78243fdb9906: Pulling fs layer 2025-09-07T07:37:02.3866505Z 68c20f3c23bb: Waiting 2025-09-07T07:37:02.3866670Z 6f70d5d50aba: Pulling fs layer 2025-09-07T07:37:02.3866854Z 4f4fb700ef54: Pulling fs layer 2025-09-07T07:37:02.3867031Z 69715d3ad3c4: Pulling fs layer 2025-09-07T07:37:02.3867209Z 7ace90c063f3: Pulling fs layer 2025-09-07T07:37:02.3867371Z 7efa39950d32: Waiting 2025-09-07T07:37:02.3867538Z acbd5447dd14: Pulling fs layer 2025-09-07T07:37:02.3867711Z 744523d9b7f5: Pulling fs layer 2025-09-07T07:37:02.3867885Z 5bd615a7b945: Pulling fs layer 2025-09-07T07:37:02.3868058Z f4986a00e3ae: Pulling fs layer 2025-09-07T07:37:02.3868242Z 21902f6e4f8c: Pulling fs layer 2025-09-07T07:37:02.3868496Z d80602abf3cc: Pulling fs layer 2025-09-07T07:37:02.3868676Z 3c51bf0bc362: Pulling fs layer 2025-09-07T07:37:02.3868847Z 119ab3bceafa: Pulling fs layer 2025-09-07T07:37:02.3869033Z af8eadc9eaab: Pulling fs layer 2025-09-07T07:37:02.3869219Z e7769b0d7a82: Pulling fs layer 2025-09-07T07:37:02.3869405Z ba263639b0f4: Pulling fs layer 2025-09-07T07:37:02.3869571Z a5ab7a280382: Pulling fs layer 2025-09-07T07:37:02.3869746Z 80b2232d952f: Pulling fs layer 2025-09-07T07:37:02.3869963Z cc93cd65e90f: Pulling fs layer 2025-09-07T07:37:02.3870131Z 1c35b7d4b67c: Waiting 2025-09-07T07:37:02.3870289Z 0eed4c15712b: Pulling fs layer 2025-09-07T07:37:02.3870452Z 092516f71fe3: Pulling fs layer 2025-09-07T07:37:02.3870622Z 8c0825014a62: Pulling fs layer 2025-09-07T07:37:02.3870793Z 8e0d2f63da0a: Pulling fs layer 2025-09-07T07:37:02.3870967Z 73aae7958ba1: Pulling fs layer 2025-09-07T07:37:02.3871135Z ac6077ec9fa5: Pulling fs layer 2025-09-07T07:37:02.3871310Z bf4ee4e45e92: Pulling fs layer 2025-09-07T07:37:02.3871486Z a10eb16a7271: Waiting 2025-09-07T07:37:02.3871642Z 3c51bf0bc362: Waiting 2025-09-07T07:37:02.3871787Z 69715d3ad3c4: Waiting 2025-09-07T07:37:02.3871950Z c1b766f9b961: Pulling fs layer 2025-09-07T07:37:02.3872124Z 6e726ef07b5d: Pulling fs layer 2025-09-07T07:37:02.3872476Z 364070434a64: Pulling fs layer 2025-09-07T07:37:02.3872642Z 71f708151a84: Pulling fs layer 2025-09-07T07:37:02.3872813Z 7ace90c063f3: Waiting 2025-09-07T07:37:02.3872966Z 7d52cf579654: Waiting 2025-09-07T07:37:02.3873122Z cc93cd65e90f: Waiting 2025-09-07T07:37:02.3873268Z 119ab3bceafa: Waiting 2025-09-07T07:37:02.3873426Z 622d8cfb39ea: Pulling fs layer 2025-09-07T07:37:02.3873596Z 0eed4c15712b: Waiting 2025-09-07T07:37:02.3873753Z af8eadc9eaab: Waiting 2025-09-07T07:37:02.3873907Z 284119a92cb1: Pulling fs layer 2025-09-07T07:37:02.3874077Z cb6a20fcf4e2: Waiting 2025-09-07T07:37:02.3874232Z 092516f71fe3: Waiting 2025-09-07T07:37:02.3874380Z e7769b0d7a82: Waiting 2025-09-07T07:37:02.3874519Z 8c0825014a62: Waiting 2025-09-07T07:37:02.3874744Z 96695940d842: Pulling fs layer 2025-09-07T07:37:02.3874913Z ba263639b0f4: Waiting 2025-09-07T07:37:02.3875058Z acbd5447dd14: Waiting 2025-09-07T07:37:02.3875221Z 7ddca6c4c050: Pulling fs layer 2025-09-07T07:37:02.3875392Z 744523d9b7f5: Waiting 2025-09-07T07:37:02.3875556Z a95e1f2f1aad: Pulling fs layer 2025-09-07T07:37:02.3875717Z a5ab7a280382: Waiting 2025-09-07T07:37:02.3875880Z 8085756b0cc0: Pulling fs layer 2025-09-07T07:37:02.3876048Z 8e0d2f63da0a: Waiting 2025-09-07T07:37:02.3876205Z 80b2232d952f: Waiting 2025-09-07T07:37:02.3876347Z f4986a00e3ae: Waiting 2025-09-07T07:37:02.3876510Z 7e9ff0c6f103: Pulling fs layer 2025-09-07T07:37:02.3876681Z 5bd615a7b945: Waiting 2025-09-07T07:37:02.3876838Z a625cbbc05b9: Pulling fs layer 2025-09-07T07:37:02.3876999Z 73aae7958ba1: Waiting 2025-09-07T07:37:02.3877153Z d80602abf3cc: Waiting 2025-09-07T07:37:02.3877305Z ac6077ec9fa5: Waiting 2025-09-07T07:37:02.3877456Z bf4ee4e45e92: Waiting 2025-09-07T07:37:02.3877604Z 4e2848642431: Pulling fs layer 2025-09-07T07:37:02.3877772Z 21902f6e4f8c: Waiting 2025-09-07T07:37:02.3877923Z c1b766f9b961: Waiting 2025-09-07T07:37:02.3878073Z 46fb6a8b3e1d: Waiting 2025-09-07T07:37:02.3878218Z 6e726ef07b5d: Waiting 2025-09-07T07:37:02.3878374Z 5e944f1ed1be: Pulling fs layer 2025-09-07T07:37:02.3878544Z 622d8cfb39ea: Waiting 2025-09-07T07:37:02.3878693Z 284119a92cb1: Waiting 2025-09-07T07:37:02.3878838Z 41619248f604: Pulling fs layer 2025-09-07T07:37:02.3879003Z 5ad6977cc38e: Waiting 2025-09-07T07:37:02.3879162Z be86f8c4f654: Pulling fs layer 2025-09-07T07:37:02.3879331Z a95e1f2f1aad: Waiting 2025-09-07T07:37:02.3879484Z ef1340e22a4b: Pulling fs layer 2025-09-07T07:37:02.3879652Z da63046995a2: Waiting 2025-09-07T07:37:02.3879801Z 364070434a64: Waiting 2025-09-07T07:37:02.3879957Z da8d8b696333: Pulling fs layer 2025-09-07T07:37:02.3880118Z 71f708151a84: Waiting 2025-09-07T07:37:02.3880266Z 78243fdb9906: Waiting 2025-09-07T07:37:02.3880422Z 386b0c49c498: Pulling fs layer 2025-09-07T07:37:02.3880593Z 8085756b0cc0: Waiting 2025-09-07T07:37:02.3880744Z 2b1d0ea7efe0: Pulling fs layer 2025-09-07T07:37:02.3880913Z 96695940d842: Waiting 2025-09-07T07:37:02.3881070Z 04c04be7408f: Pulling fs layer 2025-09-07T07:37:02.3881232Z 7ddca6c4c050: Waiting 2025-09-07T07:37:02.3881385Z 4e2848642431: Waiting 2025-09-07T07:37:02.3881547Z f8690caa3ac5: Pulling fs layer 2025-09-07T07:37:02.3881723Z 2908d6baaa6b: Pulling fs layer 2025-09-07T07:37:02.3881889Z 37e2336101eb: Pulling fs layer 2025-09-07T07:37:02.3882054Z 5e944f1ed1be: Waiting 2025-09-07T07:37:02.3882209Z f1ac881fde33: Pulling fs layer 2025-09-07T07:37:02.3882380Z 41619248f604: Waiting 2025-09-07T07:37:02.3882524Z a625cbbc05b9: Waiting 2025-09-07T07:37:02.3882676Z 7e9ff0c6f103: Waiting 2025-09-07T07:37:02.3882835Z 43b14c67347e: Pulling fs layer 2025-09-07T07:37:02.3883003Z be86f8c4f654: Waiting 2025-09-07T07:37:02.3883144Z da8d8b696333: Waiting 2025-09-07T07:37:02.3883294Z 386b0c49c498: Waiting 2025-09-07T07:37:02.3883448Z f8690caa3ac5: Waiting 2025-09-07T07:37:02.3883601Z 2b1d0ea7efe0: Waiting 2025-09-07T07:37:02.3883748Z ef1340e22a4b: Waiting 2025-09-07T07:37:02.3883901Z 6f70d5d50aba: Waiting 2025-09-07T07:37:02.3884051Z 2908d6baaa6b: Waiting 2025-09-07T07:37:02.3884202Z 37e2336101eb: Waiting 2025-09-07T07:37:02.3884346Z 4f4fb700ef54: Waiting 2025-09-07T07:37:02.3884545Z 04c04be7408f: Waiting 2025-09-07T07:37:02.3884702Z 43b14c67347e: Waiting 2025-09-07T07:37:02.3884859Z f1ac881fde33: Waiting 2025-09-07T07:37:02.4613131Z 18a5ee5b0e2e: Verifying Checksum 2025-09-07T07:37:02.4614640Z 18a5ee5b0e2e: Download complete 2025-09-07T07:37:02.5371388Z 1c35b7d4b67c: Verifying Checksum 2025-09-07T07:37:02.5372673Z 1c35b7d4b67c: Download complete 2025-09-07T07:37:02.6207730Z 68c20f3c23bb: Verifying Checksum 2025-09-07T07:37:02.6208079Z 68c20f3c23bb: Download complete 2025-09-07T07:37:02.6980208Z 7efa39950d32: Verifying Checksum 2025-09-07T07:37:02.6980549Z 7efa39950d32: Download complete 2025-09-07T07:37:02.7341188Z e6fdc8487bfe: Download complete 2025-09-07T07:37:02.7764953Z a10eb16a7271: Download complete 2025-09-07T07:37:02.7954958Z 7d52cf579654: Verifying Checksum 2025-09-07T07:37:02.7955295Z 7d52cf579654: Download complete 2025-09-07T07:37:02.8729561Z 46fb6a8b3e1d: Verifying Checksum 2025-09-07T07:37:02.8729943Z 46fb6a8b3e1d: Download complete 2025-09-07T07:37:02.9466918Z 5ad6977cc38e: Verifying Checksum 2025-09-07T07:37:02.9467276Z 5ad6977cc38e: Download complete 2025-09-07T07:37:03.0464603Z da63046995a2: Download complete 2025-09-07T07:37:03.1110113Z 78243fdb9906: Verifying Checksum 2025-09-07T07:37:03.1110466Z 78243fdb9906: Download complete 2025-09-07T07:37:03.3726231Z e6fdc8487bfe: Pull complete 2025-09-07T07:37:03.3932217Z 18a5ee5b0e2e: Pull complete 2025-09-07T07:37:03.9228325Z cb6a20fcf4e2: Verifying Checksum 2025-09-07T07:37:03.9228584Z cb6a20fcf4e2: Download complete 2025-09-07T07:37:03.9298689Z 4f4fb700ef54: Verifying Checksum 2025-09-07T07:37:03.9298969Z 4f4fb700ef54: Download complete 2025-09-07T07:37:04.0098370Z 69715d3ad3c4: Download complete 2025-09-07T07:37:04.0859487Z 7ace90c063f3: Verifying Checksum 2025-09-07T07:37:04.0859765Z 7ace90c063f3: Download complete 2025-09-07T07:37:04.1530282Z acbd5447dd14: Verifying Checksum 2025-09-07T07:37:04.1530564Z acbd5447dd14: Download complete 2025-09-07T07:37:04.2580887Z 744523d9b7f5: Verifying Checksum 2025-09-07T07:37:04.2581189Z 744523d9b7f5: Download complete 2025-09-07T07:37:04.3313952Z 5bd615a7b945: Verifying Checksum 2025-09-07T07:37:04.3314320Z 5bd615a7b945: Download complete 2025-09-07T07:37:04.4166752Z f4986a00e3ae: Verifying Checksum 2025-09-07T07:37:04.4167100Z f4986a00e3ae: Download complete 2025-09-07T07:37:04.5086276Z 21902f6e4f8c: Download complete 2025-09-07T07:37:04.5841891Z d80602abf3cc: Verifying Checksum 2025-09-07T07:37:04.5842148Z d80602abf3cc: Download complete 2025-09-07T07:37:04.6688245Z 3c51bf0bc362: Download complete 2025-09-07T07:37:04.7592825Z 119ab3bceafa: Verifying Checksum 2025-09-07T07:37:04.7593078Z 119ab3bceafa: Download complete 2025-09-07T07:37:04.8387876Z af8eadc9eaab: Verifying Checksum 2025-09-07T07:37:04.8388144Z af8eadc9eaab: Download complete 2025-09-07T07:37:04.9176439Z e7769b0d7a82: Download complete 2025-09-07T07:37:05.5831254Z 572424b92528: Verifying Checksum 2025-09-07T07:37:05.5831497Z 572424b92528: Download complete 2025-09-07T07:37:05.6549133Z a5ab7a280382: Verifying Checksum 2025-09-07T07:37:05.6549464Z a5ab7a280382: Download complete 2025-09-07T07:37:05.7530778Z 80b2232d952f: Verifying Checksum 2025-09-07T07:37:05.7531116Z 80b2232d952f: Download complete 2025-09-07T07:37:05.8306136Z cc93cd65e90f: Verifying Checksum 2025-09-07T07:37:05.8306381Z cc93cd65e90f: Download complete 2025-09-07T07:37:05.9241806Z 0eed4c15712b: Verifying Checksum 2025-09-07T07:37:05.9242159Z 0eed4c15712b: Download complete 2025-09-07T07:37:06.1783945Z 092516f71fe3: Verifying Checksum 2025-09-07T07:37:06.1784217Z 092516f71fe3: Download complete 2025-09-07T07:37:06.2417037Z 8c0825014a62: Download complete 2025-09-07T07:37:06.3101694Z 8e0d2f63da0a: Verifying Checksum 2025-09-07T07:37:06.3102089Z 8e0d2f63da0a: Download complete 2025-09-07T07:37:06.3990945Z 73aae7958ba1: Verifying Checksum 2025-09-07T07:37:06.3991306Z 73aae7958ba1: Download complete 2025-09-07T07:37:06.4794756Z ac6077ec9fa5: Download complete 2025-09-07T07:37:06.5518499Z bf4ee4e45e92: Verifying Checksum 2025-09-07T07:37:06.5520008Z bf4ee4e45e92: Download complete 2025-09-07T07:37:09.5167499Z ba263639b0f4: Verifying Checksum 2025-09-07T07:37:09.5167853Z ba263639b0f4: Download complete 2025-09-07T07:37:09.6420711Z 6e726ef07b5d: Verifying Checksum 2025-09-07T07:37:09.6421987Z 6e726ef07b5d: Download complete 2025-09-07T07:37:11.5386917Z 572424b92528: Pull complete 2025-09-07T07:37:12.0547921Z 1c35b7d4b67c: Pull complete 2025-09-07T07:37:12.4807985Z 68c20f3c23bb: Pull complete 2025-09-07T07:37:12.4857365Z 364070434a64: Verifying Checksum 2025-09-07T07:37:12.4857575Z 364070434a64: Download complete 2025-09-07T07:37:12.9269563Z 7efa39950d32: Pull complete 2025-09-07T07:37:13.4096041Z a10eb16a7271: Pull complete 2025-09-07T07:37:13.8500402Z 7d52cf579654: Pull complete 2025-09-07T07:37:16.0441057Z cb6a20fcf4e2: Pull complete 2025-09-07T07:37:16.5244142Z 46fb6a8b3e1d: Pull complete 2025-09-07T07:37:16.8543799Z 5ad6977cc38e: Pull complete 2025-09-07T07:37:17.1558475Z da63046995a2: Pull complete 2025-09-07T07:37:17.4150336Z 78243fdb9906: Pull complete 2025-09-07T07:37:37.1127582Z 6f70d5d50aba: Verifying Checksum 2025-09-07T07:37:37.1127910Z 6f70d5d50aba: Download complete 2025-09-07T07:37:37.1974131Z 622d8cfb39ea: Download complete 2025-09-07T07:37:37.2695909Z 284119a92cb1: Verifying Checksum 2025-09-07T07:37:37.2696166Z 284119a92cb1: Download complete 2025-09-07T07:37:37.3629730Z 96695940d842: Download complete 2025-09-07T07:37:37.4856438Z 7ddca6c4c050: Verifying Checksum 2025-09-07T07:37:37.4858111Z 7ddca6c4c050: Download complete 2025-09-07T07:37:37.5749132Z a95e1f2f1aad: Verifying Checksum 2025-09-07T07:37:37.5751101Z a95e1f2f1aad: Download complete 2025-09-07T07:37:37.6527925Z 8085756b0cc0: Download complete 2025-09-07T07:37:37.7388140Z 7e9ff0c6f103: Download complete 2025-09-07T07:37:37.8178115Z a625cbbc05b9: Verifying Checksum 2025-09-07T07:37:37.8178376Z a625cbbc05b9: Download complete 2025-09-07T07:37:37.8971713Z 4e2848642431: Download complete 2025-09-07T07:37:37.9830195Z 5e944f1ed1be: Download complete 2025-09-07T07:37:38.0651578Z 41619248f604: Download complete 2025-09-07T07:37:38.1343156Z be86f8c4f654: Download complete 2025-09-07T07:37:38.2360343Z ef1340e22a4b: Download complete 2025-09-07T07:37:38.2877982Z da8d8b696333: Download complete 2025-09-07T07:37:40.7821425Z 386b0c49c498: Verifying Checksum 2025-09-07T07:37:40.7821719Z 386b0c49c498: Download complete 2025-09-07T07:37:40.8896967Z 2b1d0ea7efe0: Verifying Checksum 2025-09-07T07:37:40.8897236Z 2b1d0ea7efe0: Download complete 2025-09-07T07:37:40.9805473Z 04c04be7408f: Download complete 2025-09-07T07:37:41.0643206Z f8690caa3ac5: Verifying Checksum 2025-09-07T07:37:41.0643432Z f8690caa3ac5: Download complete 2025-09-07T07:37:41.1508419Z 2908d6baaa6b: Verifying Checksum 2025-09-07T07:37:41.1508701Z 2908d6baaa6b: Download complete 2025-09-07T07:37:41.2112076Z 37e2336101eb: Download complete 2025-09-07T07:37:41.3136941Z f1ac881fde33: Download complete 2025-09-07T07:37:41.9150970Z 43b14c67347e: Verifying Checksum 2025-09-07T07:37:41.9151260Z 43b14c67347e: Download complete 2025-09-07T07:38:12.8959899Z 6f70d5d50aba: Pull complete 2025-09-07T07:38:13.1278537Z 4f4fb700ef54: Pull complete 2025-09-07T07:38:13.4084230Z 69715d3ad3c4: Pull complete 2025-09-07T07:38:13.6675180Z 7ace90c063f3: Pull complete 2025-09-07T07:38:13.9539432Z acbd5447dd14: Pull complete 2025-09-07T07:38:14.5528527Z 744523d9b7f5: Pull complete 2025-09-07T07:38:15.0038636Z 5bd615a7b945: Pull complete 2025-09-07T07:38:15.4196952Z f4986a00e3ae: Pull complete 2025-09-07T07:38:15.9221990Z 21902f6e4f8c: Pull complete 2025-09-07T07:38:16.4218319Z d80602abf3cc: Pull complete 2025-09-07T07:38:16.9635384Z 3c51bf0bc362: Pull complete 2025-09-07T07:38:17.0562115Z 71f708151a84: Verifying Checksum 2025-09-07T07:38:17.0562371Z 71f708151a84: Download complete 2025-09-07T07:38:17.5174568Z 119ab3bceafa: Pull complete 2025-09-07T07:38:18.4574180Z af8eadc9eaab: Pull complete 2025-09-07T07:38:18.8098206Z e7769b0d7a82: Pull complete 2025-09-07T07:38:24.8176728Z ba263639b0f4: Pull complete 2025-09-07T07:38:25.2767981Z a5ab7a280382: Pull complete 2025-09-07T07:38:25.7393902Z 80b2232d952f: Pull complete 2025-09-07T07:38:26.5731839Z cc93cd65e90f: Pull complete 2025-09-07T07:38:27.0252305Z 0eed4c15712b: Pull complete 2025-09-07T07:38:27.6186172Z 092516f71fe3: Pull complete 2025-09-07T07:38:28.0550601Z 8c0825014a62: Pull complete 2025-09-07T07:38:28.6212301Z 8e0d2f63da0a: Pull complete 2025-09-07T07:38:29.5570788Z 73aae7958ba1: Pull complete 2025-09-07T07:38:29.9523806Z ac6077ec9fa5: Pull complete 2025-09-07T07:38:30.3537995Z bf4ee4e45e92: Pull complete 2025-09-07T07:40:12.7815907Z c1b766f9b961: Verifying Checksum 2025-09-07T07:40:12.7816208Z c1b766f9b961: Download complete 2025-09-07T07:42:14.5440965Z c1b766f9b961: Pull complete 2025-09-07T07:42:14.8609318Z 6e726ef07b5d: Pull complete 2025-09-07T07:42:16.3315250Z 364070434a64: Pull complete 2025-09-07T07:43:26.6904502Z 71f708151a84: Pull complete 2025-09-07T07:43:27.0299829Z 622d8cfb39ea: Pull complete 2025-09-07T07:43:27.2903230Z 284119a92cb1: Pull complete 2025-09-07T07:43:27.8770176Z 96695940d842: Pull complete 2025-09-07T07:43:28.4797869Z 7ddca6c4c050: Pull complete 2025-09-07T07:43:28.7468335Z a95e1f2f1aad: Pull complete 2025-09-07T07:43:29.4506037Z 8085756b0cc0: Pull complete 2025-09-07T07:43:30.2332096Z 7e9ff0c6f103: Pull complete 2025-09-07T07:43:30.7039162Z a625cbbc05b9: Pull complete 2025-09-07T07:43:31.5623985Z 4e2848642431: Pull complete 2025-09-07T07:43:31.9938110Z 5e944f1ed1be: Pull complete 2025-09-07T07:43:32.7892769Z 41619248f604: Pull complete 2025-09-07T07:43:33.3317823Z be86f8c4f654: Pull complete 2025-09-07T07:43:34.2834694Z ef1340e22a4b: Pull complete 2025-09-07T07:43:34.6199335Z da8d8b696333: Pull complete 2025-09-07T07:43:39.6609619Z 386b0c49c498: Pull complete 2025-09-07T07:43:39.9124958Z 2b1d0ea7efe0: Pull complete 2025-09-07T07:43:40.3250648Z 04c04be7408f: Pull complete 2025-09-07T07:43:40.6029149Z f8690caa3ac5: Pull complete 2025-09-07T07:43:40.8429486Z 2908d6baaa6b: Pull complete 2025-09-07T07:43:41.3178616Z 37e2336101eb: Pull complete 2025-09-07T07:43:41.9022463Z f1ac881fde33: Pull complete 2025-09-07T07:43:43.3796134Z 43b14c67347e: Pull complete 2025-09-07T07:43:44.0511213Z Digest: sha256:383efb45082f20b8c808cb0ba4df693a01359592233f641f1f486911ac320a9a 2025-09-07T07:43:44.1311571Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:43:44.1654605Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:43:44.1725356Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:43:44.1725988Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:43:44.1734456Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:44.1734730Z env: 2025-09-07T07:43:44.1734891Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:44.1735072Z ##[endgroup] 2025-09-07T07:43:44.1799957Z Prepare all required actions 2025-09-07T07:43:44.1847909Z ##[group]Run ./.github/actions/get-workflow-job-id 2025-09-07T07:43:44.1848125Z with: 2025-09-07T07:43:44.1848619Z github-token: *** 2025-09-07T07:43:44.1848784Z env: 2025-09-07T07:43:44.1848934Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:44.1849104Z ##[endgroup] 2025-09-07T07:43:44.1945326Z ##[group]Run set -eux 2025-09-07T07:43:44.1945501Z set -eux 2025-09-07T07:43:44.1945771Z python3 .github/scripts/get_workflow_job_id.py "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-09-07T07:43:44.1949848Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:44.1950079Z env: 2025-09-07T07:43:44.1950239Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:44.1950557Z GITHUB_TOKEN: *** 2025-09-07T07:43:44.1950723Z ##[endgroup] 2025-09-07T07:43:44.1967383Z + python3 .github/scripts/get_workflow_job_id.py 17525285611 i-05ba5cf17983f662f 2025-09-07T07:43:44.8196904Z Setting output job-id=49775585800 2025-09-07T07:43:44.8197477Z Setting output job-name=inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal) 2025-09-07T07:43:44.8287971Z ##[group]Run python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-09-07T07:43:44.8288396Z python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-09-07T07:43:44.8288927Z python3 -m tools.stats.monitor --log-interval "$MONITOR_LOG_INTERVAL" --data-collect-interval "$MONITOR_DATA_COLLECT_INTERVAL" > usage_log.txt 2>&1 & 2025-09-07T07:43:44.8289399Z echo "monitor-script-pid=${!}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:43:44.8293434Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:44.8293657Z env: 2025-09-07T07:43:44.8293807Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:44.8293978Z JOB_ID: 49775585800 2025-09-07T07:43:44.8294301Z JOB_NAME: inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal) 2025-09-07T07:43:44.8294669Z WORKFLOW_NAME: inductor-perf-nightly-x86 2025-09-07T07:43:44.8294906Z WORKFLOW_RUN_ID: 17525285611 2025-09-07T07:43:44.8295078Z MONITOR_LOG_INTERVAL: 15 2025-09-07T07:43:44.8295253Z MONITOR_DATA_COLLECT_INTERVAL: 4 2025-09-07T07:43:44.8295439Z ##[endgroup] 2025-09-07T07:43:45.3625087Z Defaulting to user installation because normal site-packages is not writeable 2025-09-07T07:43:45.7754886Z Collecting psutil==5.9.8 2025-09-07T07:43:45.7883019Z Downloading psutil-5.9.8-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (288 kB) 2025-09-07T07:43:45.9381626Z Collecting dataclasses_json==0.6.7 2025-09-07T07:43:45.9409569Z Downloading dataclasses_json-0.6.7-py3-none-any.whl (28 kB) 2025-09-07T07:43:46.0847270Z Collecting nvidia-ml-py==11.525.84 2025-09-07T07:43:46.0882891Z Downloading nvidia_ml_py-11.525.84-py3-none-any.whl (34 kB) 2025-09-07T07:43:46.1802485Z Collecting marshmallow<4.0.0,>=3.18.0 2025-09-07T07:43:46.1831021Z Downloading marshmallow-3.26.1-py3-none-any.whl (50 kB) 2025-09-07T07:43:46.2447593Z Collecting typing-inspect<1,>=0.4.0 2025-09-07T07:43:46.2476471Z Downloading typing_inspect-0.9.0-py3-none-any.whl (8.8 kB) 2025-09-07T07:43:46.3438811Z Collecting packaging>=17.0 2025-09-07T07:43:46.3466825Z Downloading packaging-25.0-py3-none-any.whl (66 kB) 2025-09-07T07:43:46.4804219Z Collecting mypy-extensions>=0.3.0 2025-09-07T07:43:46.4830033Z Downloading mypy_extensions-1.1.0-py3-none-any.whl (5.0 kB) 2025-09-07T07:43:46.6083704Z Collecting typing-extensions>=3.7.4 2025-09-07T07:43:46.6113477Z Downloading typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-09-07T07:43:46.8261924Z Installing collected packages: typing-extensions, packaging, mypy-extensions, typing-inspect, marshmallow, psutil, nvidia-ml-py, dataclasses-json 2025-09-07T07:43:47.5217102Z Successfully installed dataclasses-json-0.6.7 marshmallow-3.26.1 mypy-extensions-1.1.0 nvidia-ml-py-11.525.84 packaging-25.0 psutil-5.9.8 typing-extensions-4.15.0 typing-inspect-0.9.0 2025-09-07T07:43:47.7139732Z Prepare all required actions 2025-09-07T07:43:47.7139998Z Getting action download info 2025-09-07T07:43:47.9641447Z Download action repository 'seemethere/download-artifact-s3@v4' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-09-07T07:43:48.3404327Z Download action repository 'actions/download-artifact@v4' (SHA:d3f86a106a0bac45b974a628896c90dbdf5c8093) 2025-09-07T07:43:48.6875499Z ##[group]Run ./.github/actions/download-build-artifacts 2025-09-07T07:43:48.6875727Z with: 2025-09-07T07:43:48.6875904Z name: linux-jammy-py3.9-gcc11-build 2025-09-07T07:43:48.6876109Z s3-bucket: gha-artifacts 2025-09-07T07:43:48.6876287Z env: 2025-09-07T07:43:48.6876443Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:48.6876612Z ##[endgroup] 2025-09-07T07:43:48.7002144Z ##[group]Run seemethere/download-artifact-s3@v4 2025-09-07T07:43:48.7002468Z with: 2025-09-07T07:43:48.7002637Z name: linux-jammy-py3.9-gcc11-build 2025-09-07T07:43:48.7002837Z s3-bucket: gha-artifacts 2025-09-07T07:43:48.7003050Z region: us-east-1 2025-09-07T07:43:48.7003192Z env: 2025-09-07T07:43:48.7003333Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:48.7003502Z ##[endgroup] 2025-09-07T07:43:49.3061943Z (node:57942) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-09-07T07:43:49.3062341Z 2025-09-07T07:43:49.3062549Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-09-07T07:43:49.3062874Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-09-07T07:43:49.3063232Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-09-07T07:43:50.5266304Z Found 1 objects with prefix pytorch/pytorch/17525285611/linux-jammy-py3.9-gcc11-build/ 2025-09-07T07:43:50.5266757Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-09-07T07:43:55.1390870Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-09-07T07:43:55.1393792Z Artifact download has finished successfully 2025-09-07T07:43:55.1585060Z ##[group]Run unzip -o artifacts.zip 2025-09-07T07:43:55.1585290Z unzip -o artifacts.zip 2025-09-07T07:43:55.1590596Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:55.1590819Z env: 2025-09-07T07:43:55.1590972Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:55.1591150Z ##[endgroup] 2025-09-07T07:43:55.2114604Z Archive: artifacts.zip 2025-09-07T07:43:55.2114831Z creating: dist/ 2025-09-07T07:43:56.2223892Z inflating: dist/torch-2.9.0a0+git93fb23d-cp39-cp39-linux_x86_64.whl 2025-09-07T07:43:56.2224295Z creating: dist/vision/ 2025-09-07T07:43:56.2294658Z inflating: dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl 2025-09-07T07:43:56.2294949Z creating: dist/audio/ 2025-09-07T07:43:56.2322441Z inflating: dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl 2025-09-07T07:43:56.2322819Z creating: dist/ao/ 2025-09-07T07:43:56.2357370Z inflating: dist/ao/torchao-0.7.0+git51c87b6e-py3-none-any.whl 2025-09-07T07:43:56.2463078Z inflating: dist/.ninja_log 2025-09-07T07:43:56.2463342Z creating: build/custom_test_artifacts/ 2025-09-07T07:43:56.2463604Z creating: build/custom_test_artifacts/custom-op-build/ 2025-09-07T07:43:56.2463905Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/ 2025-09-07T07:43:56.2464248Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/pkgRedirects/ 2025-09-07T07:43:56.2465988Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeConfigureLog.yaml 2025-09-07T07:43:56.2466362Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/ 2025-09-07T07:43:56.2466728Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeSystem.cmake 2025-09-07T07:43:56.2467119Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/ 2025-09-07T07:43:56.2467764Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/tmp/ 2025-09-07T07:43:56.2468691Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/CMakeCCompilerId.c 2025-09-07T07:43:56.2469692Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/a.out 2025-09-07T07:43:56.2470120Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeCCompiler.cmake 2025-09-07T07:43:56.2470517Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/ 2025-09-07T07:43:56.2470903Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/tmp/ 2025-09-07T07:43:56.2472390Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-09-07T07:43:56.2473412Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/a.out 2025-09-07T07:43:56.2474026Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeCXXCompiler.cmake 2025-09-07T07:43:56.2475072Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_C.bin 2025-09-07T07:43:56.2476406Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_CXX.bin 2025-09-07T07:43:56.2476830Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeScratch/ 2025-09-07T07:43:56.2477194Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/cmake.check_cache 2025-09-07T07:43:56.2477560Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/ 2025-09-07T07:43:56.2477973Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.ts 2025-09-07T07:43:56.2478436Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.make 2025-09-07T07:43:56.2478882Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/depend.make 2025-09-07T07:43:56.2479295Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.txt 2025-09-07T07:43:56.2479734Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/cmake_clean.cmake 2025-09-07T07:43:56.2480160Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/build.make 2025-09-07T07:43:56.2480588Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/DependInfo.cmake 2025-09-07T07:43:56.2481041Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/flags.make 2025-09-07T07:43:56.2481500Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/progress.make 2025-09-07T07:43:56.2496084Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o.d 2025-09-07T07:43:56.2658113Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o 2025-09-07T07:43:56.2658505Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/ 2025-09-07T07:43:56.2658938Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.ts 2025-09-07T07:43:56.2660496Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.make 2025-09-07T07:43:56.2660960Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/depend.make 2025-09-07T07:43:56.2661387Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.txt 2025-09-07T07:43:56.2661821Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/cmake_clean.cmake 2025-09-07T07:43:56.2662260Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/build.make 2025-09-07T07:43:56.2662797Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/DependInfo.cmake 2025-09-07T07:43:56.2663246Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/flags.make 2025-09-07T07:43:56.2663678Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/progress.make 2025-09-07T07:43:56.2677649Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o.d 2025-09-07T07:43:56.2743721Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o 2025-09-07T07:43:56.2744259Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-09-07T07:43:56.2744696Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/TargetDirectories.txt 2025-09-07T07:43:56.2745154Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/progress.marks 2025-09-07T07:43:56.2745529Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile2 2025-09-07T07:43:56.2746348Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile.cmake 2025-09-07T07:43:56.2746739Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/InstallScripts.json 2025-09-07T07:43:56.2747654Z inflating: build/custom_test_artifacts/custom-op-build/CMakeCache.txt 2025-09-07T07:43:56.2748236Z inflating: build/custom_test_artifacts/custom-op-build/Makefile 2025-09-07T07:43:56.2748621Z inflating: build/custom_test_artifacts/custom-op-build/cmake_install.cmake 2025-09-07T07:43:56.2892614Z inflating: build/custom_test_artifacts/custom-op-build/libcustom_ops.so 2025-09-07T07:43:56.2938132Z inflating: build/custom_test_artifacts/custom-op-build/test_custom_ops 2025-09-07T07:43:56.2938486Z creating: build/custom_test_artifacts/jit-hook-build/ 2025-09-07T07:43:56.2938824Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/ 2025-09-07T07:43:56.2939194Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/pkgRedirects/ 2025-09-07T07:43:56.2941583Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeConfigureLog.yaml 2025-09-07T07:43:56.2941952Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/ 2025-09-07T07:43:56.2942315Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeSystem.cmake 2025-09-07T07:43:56.2942712Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/ 2025-09-07T07:43:56.2943093Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/tmp/ 2025-09-07T07:43:56.2944234Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/CMakeCCompilerId.c 2025-09-07T07:43:56.2945218Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/a.out 2025-09-07T07:43:56.2945919Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeCCompiler.cmake 2025-09-07T07:43:56.2946389Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/ 2025-09-07T07:43:56.2946793Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/tmp/ 2025-09-07T07:43:56.2947985Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-09-07T07:43:56.2948964Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/a.out 2025-09-07T07:43:56.2949566Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeCXXCompiler.cmake 2025-09-07T07:43:56.2950618Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_C.bin 2025-09-07T07:43:56.2951756Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_CXX.bin 2025-09-07T07:43:56.2952195Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeScratch/ 2025-09-07T07:43:56.2952756Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/cmake.check_cache 2025-09-07T07:43:56.2953139Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/ 2025-09-07T07:43:56.2953571Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.ts 2025-09-07T07:43:56.2954036Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.make 2025-09-07T07:43:56.2954489Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/depend.make 2025-09-07T07:43:56.2954918Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.txt 2025-09-07T07:43:56.2955356Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/cmake_clean.cmake 2025-09-07T07:43:56.2955869Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/build.make 2025-09-07T07:43:56.2956309Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/DependInfo.cmake 2025-09-07T07:43:56.2956743Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/flags.make 2025-09-07T07:43:56.2957178Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/progress.make 2025-09-07T07:43:56.2971566Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o.d 2025-09-07T07:43:56.3023425Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o 2025-09-07T07:43:56.3023903Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-09-07T07:43:56.3024332Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/TargetDirectories.txt 2025-09-07T07:43:56.3024733Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/progress.marks 2025-09-07T07:43:56.3025104Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile2 2025-09-07T07:43:56.3025652Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile.cmake 2025-09-07T07:43:56.3026274Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/InstallScripts.json 2025-09-07T07:43:56.3027353Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeCache.txt 2025-09-07T07:43:56.3027828Z inflating: build/custom_test_artifacts/jit-hook-build/Makefile 2025-09-07T07:43:56.3028247Z inflating: build/custom_test_artifacts/jit-hook-build/cmake_install.cmake 2025-09-07T07:43:56.3059980Z inflating: build/custom_test_artifacts/jit-hook-build/test_jit_hooks 2025-09-07T07:43:56.3060290Z creating: build/custom_test_artifacts/custom-backend-build/ 2025-09-07T07:43:56.3060600Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/ 2025-09-07T07:43:56.3060979Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/pkgRedirects/ 2025-09-07T07:43:56.3063465Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeConfigureLog.yaml 2025-09-07T07:43:56.3063872Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/ 2025-09-07T07:43:56.3064259Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeSystem.cmake 2025-09-07T07:43:56.3064676Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/ 2025-09-07T07:43:56.3065084Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/tmp/ 2025-09-07T07:43:56.3066066Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/CMakeCCompilerId.c 2025-09-07T07:43:56.3067052Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/a.out 2025-09-07T07:43:56.3067649Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeCCompiler.cmake 2025-09-07T07:43:56.3068090Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/ 2025-09-07T07:43:56.3068548Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/tmp/ 2025-09-07T07:43:56.3069850Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-09-07T07:43:56.3070844Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/a.out 2025-09-07T07:43:56.3071404Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeCXXCompiler.cmake 2025-09-07T07:43:56.3072553Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_C.bin 2025-09-07T07:43:56.3073659Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_CXX.bin 2025-09-07T07:43:56.3074174Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeScratch/ 2025-09-07T07:43:56.3074565Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/cmake.check_cache 2025-09-07T07:43:56.3074960Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/ 2025-09-07T07:43:56.3075412Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.ts 2025-09-07T07:43:56.3075910Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.make 2025-09-07T07:43:56.3076386Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/depend.make 2025-09-07T07:43:56.3076840Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.txt 2025-09-07T07:43:56.3077295Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/cmake_clean.cmake 2025-09-07T07:43:56.3077771Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/build.make 2025-09-07T07:43:56.3078235Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/DependInfo.cmake 2025-09-07T07:43:56.3078695Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/flags.make 2025-09-07T07:43:56.3079151Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/progress.make 2025-09-07T07:43:56.3080343Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o.d 2025-09-07T07:43:56.3177864Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o 2025-09-07T07:43:56.3178335Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/ 2025-09-07T07:43:56.3178812Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.ts 2025-09-07T07:43:56.3179336Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.make 2025-09-07T07:43:56.3179841Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/depend.make 2025-09-07T07:43:56.3180316Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.txt 2025-09-07T07:43:56.3180797Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/cmake_clean.cmake 2025-09-07T07:43:56.3181289Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/build.make 2025-09-07T07:43:56.3181785Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/DependInfo.cmake 2025-09-07T07:43:56.3182333Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/flags.make 2025-09-07T07:43:56.3182830Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/progress.make 2025-09-07T07:43:56.3197543Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o.d 2025-09-07T07:43:56.3242029Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o 2025-09-07T07:43:56.3242546Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-09-07T07:43:56.3243007Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/TargetDirectories.txt 2025-09-07T07:43:56.3243417Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/progress.marks 2025-09-07T07:43:56.3243853Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile2 2025-09-07T07:43:56.3244409Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile.cmake 2025-09-07T07:43:56.3244814Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/InstallScripts.json 2025-09-07T07:43:56.3245734Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeCache.txt 2025-09-07T07:43:56.3246318Z inflating: build/custom_test_artifacts/custom-backend-build/Makefile 2025-09-07T07:43:56.3246739Z inflating: build/custom_test_artifacts/custom-backend-build/cmake_install.cmake 2025-09-07T07:43:56.3330443Z inflating: build/custom_test_artifacts/custom-backend-build/libcustom_backend.so 2025-09-07T07:43:56.3363264Z inflating: build/custom_test_artifacts/custom-backend-build/test_custom_backend 2025-09-07T07:43:56.3363544Z creating: build/lib/ 2025-09-07T07:43:56.3433253Z inflating: build/lib/libprotobuf-lite.a 2025-09-07T07:43:56.3809210Z inflating: build/lib/libprotobuf.a 2025-09-07T07:43:56.4229564Z inflating: build/lib/libprotoc.a 2025-09-07T07:43:56.4237847Z inflating: build/lib/libpthreadpool.a 2025-09-07T07:43:56.4244299Z inflating: build/lib/libcpuinfo.a 2025-09-07T07:43:56.4250880Z inflating: build/lib/libcpuinfo_internals.a 2025-09-07T07:43:56.4251489Z inflating: build/lib/libclog.a 2025-09-07T07:43:56.4267422Z inflating: build/lib/libpytorch_qnnpack.a 2025-09-07T07:43:56.4269233Z inflating: build/lib/libnnpack_reference_layers.a 2025-09-07T07:43:56.4429457Z inflating: build/lib/libmicrokernels-prod.a 2025-09-07T07:43:56.4443116Z inflating: build/lib/libnnpack.a 2025-09-07T07:43:56.5189054Z inflating: build/lib/libmicrokernels-all.a 2025-09-07T07:43:56.5248454Z inflating: build/lib/libgtest.a 2025-09-07T07:43:56.5262901Z inflating: build/lib/libgmock.a 2025-09-07T07:43:56.5263369Z inflating: build/lib/libgtest_main.a 2025-09-07T07:43:56.5263810Z inflating: build/lib/libgmock_main.a 2025-09-07T07:43:56.5340487Z inflating: build/lib/libXNNPACK.a 2025-09-07T07:43:56.5404810Z inflating: build/lib/libbenchmark.a 2025-09-07T07:43:56.5405322Z inflating: build/lib/libbenchmark_main.a 2025-09-07T07:43:56.5405843Z inflating: build/lib/libjitprofiling.a 2025-09-07T07:43:56.5412330Z inflating: build/lib/libittnotify.a 2025-09-07T07:43:56.5468464Z inflating: build/lib/libasmjit.a 2025-09-07T07:43:56.6449789Z inflating: build/lib/libfbgemm.a 2025-09-07T07:43:56.6474953Z inflating: build/lib/libtensorpipe_uv.a 2025-09-07T07:43:56.6938680Z inflating: build/lib/libtensorpipe.a 2025-09-07T07:43:56.7039837Z inflating: build/lib/libgloo.a 2025-09-07T07:43:56.7079167Z inflating: build/lib/libonnx_proto.a 2025-09-07T07:43:56.7682251Z inflating: build/lib/libonnx.a 2025-09-07T07:43:57.6195574Z inflating: build/lib/libdnnl.a 2025-09-07T07:43:57.6209749Z inflating: build/lib/libfmt.a 2025-09-07T07:43:57.6435022Z inflating: build/lib/libkineto.a 2025-09-07T07:43:57.6528851Z inflating: build/lib/libc10.so 2025-09-07T07:43:57.6530281Z inflating: build/lib/libtorch_global_deps.so 2025-09-07T07:44:00.1993384Z inflating: build/lib/libtorch_cpu.so 2025-09-07T07:44:00.1993655Z inflating: build/lib/libtorch.so 2025-09-07T07:44:00.2053317Z inflating: build/lib/libtorchbind_test.so 2025-09-07T07:44:00.2068881Z inflating: build/lib/libjitbackend_test.so 2025-09-07T07:44:00.2089431Z inflating: build/lib/libbackend_with_compiler.so 2025-09-07T07:44:00.2111902Z inflating: build/lib/libaoti_custom_ops.so 2025-09-07T07:44:00.2114961Z inflating: build/lib/libshm.so 2025-09-07T07:44:00.3867269Z inflating: build/lib/libtorch_python.so 2025-09-07T07:44:00.3896898Z inflating: build/lib/libnnapi_backend.so 2025-09-07T07:44:00.3897173Z creating: build/bin/ 2025-09-07T07:44:00.3897367Z creating: build/bin/CMakeFiles/ 2025-09-07T07:44:00.3897669Z inflating: build/bin/cmake_install.cmake 2025-09-07T07:44:00.3898424Z inflating: build/bin/CTestTestfile.cmake 2025-09-07T07:44:00.4291393Z inflating: build/bin/protoc-3.13.0.0 2025-09-07T07:44:00.4686208Z inflating: build/bin/protoc 2025-09-07T07:44:00.4735148Z inflating: build/bin/c10_AllocatorConfig_test 2025-09-07T07:44:00.4782404Z inflating: build/bin/c10_CompileTimeFunctionPointer_test 2025-09-07T07:44:00.4831579Z inflating: build/bin/c10_DeviceGuard_test 2025-09-07T07:44:00.4880853Z inflating: build/bin/c10_Device_test 2025-09-07T07:44:00.4936658Z inflating: build/bin/c10_DispatchKeySet_test 2025-09-07T07:44:00.4983483Z inflating: build/bin/c10_StreamGuard_test 2025-09-07T07:44:00.5034671Z inflating: build/bin/c10_Scalar_test 2025-09-07T07:44:00.5088689Z inflating: build/bin/c10_SymInt_test 2025-09-07T07:44:00.5140129Z inflating: build/bin/c10_InlineDeviceGuard_test 2025-09-07T07:44:00.5193170Z inflating: build/bin/c10_SizesAndStrides_test 2025-09-07T07:44:00.5245791Z inflating: build/bin/c10_InlineStreamGuard_test 2025-09-07T07:44:00.5293227Z inflating: build/bin/c10_ArrayRef_test 2025-09-07T07:44:00.5339741Z inflating: build/bin/c10_ConstexprCrc_test 2025-09-07T07:44:00.5405319Z inflating: build/bin/c10_cow_test 2025-09-07T07:44:00.5455398Z inflating: build/bin/c10_Bitset_test 2025-09-07T07:44:00.5509263Z inflating: build/bin/c10_Enumerate_test 2025-09-07T07:44:00.5556674Z inflating: build/bin/c10_DeadlockDetection_test 2025-09-07T07:44:00.5607031Z inflating: build/bin/c10_IntrusiveList_test 2025-09-07T07:44:00.5659924Z inflating: build/bin/c10_LeftRight_test 2025-09-07T07:44:00.5708098Z inflating: build/bin/c10_Half_test 2025-09-07T07:44:00.5760719Z inflating: build/bin/c10_Metaprogramming_test 2025-09-07T07:44:00.5811426Z inflating: build/bin/c10_NetworkFlow_test 2025-09-07T07:44:00.5858756Z inflating: build/bin/c10_Semaphore_test 2025-09-07T07:44:00.5906384Z inflating: build/bin/c10_Synchronized_test 2025-09-07T07:44:00.5955773Z inflating: build/bin/c10_TypeIndex_test 2025-09-07T07:44:00.6008160Z inflating: build/bin/c10_ThreadLocal_test 2025-09-07T07:44:00.6056580Z inflating: build/bin/c10_TypeList_test 2025-09-07T07:44:00.6103449Z inflating: build/bin/c10_TypeTraits_test 2025-09-07T07:44:00.6152353Z inflating: build/bin/c10_accumulate_test 2025-09-07T07:44:00.6205356Z inflating: build/bin/c10_bfloat16_test 2025-09-07T07:44:00.6253217Z inflating: build/bin/c10_bit_cast_test 2025-09-07T07:44:00.6306826Z inflating: build/bin/c10_complex_math_test 2025-09-07T07:44:00.6358992Z inflating: build/bin/c10_complex_test 2025-09-07T07:44:00.6406232Z inflating: build/bin/c10_error_test 2025-09-07T07:44:00.6456018Z inflating: build/bin/c10_exception_test 2025-09-07T07:44:00.6503842Z inflating: build/bin/c10_flags_test 2025-09-07T07:44:00.6551768Z inflating: build/bin/c10_generic_math_test 2025-09-07T07:44:00.6600288Z inflating: build/bin/c10_irange_test 2025-09-07T07:44:00.6745509Z inflating: build/bin/c10_intrusive_ptr_test 2025-09-07T07:44:00.6796272Z inflating: build/bin/c10_lazy_test 2025-09-07T07:44:00.6850330Z inflating: build/bin/c10_logging_test 2025-09-07T07:44:00.6920207Z inflating: build/bin/c10_optional_test 2025-09-07T07:44:00.6978267Z inflating: build/bin/c10_ordered_preserving_dict_test 2025-09-07T07:44:00.7028932Z inflating: build/bin/c10_registry_test 2025-09-07T07:44:00.7166023Z inflating: build/bin/c10_small_vector_test 2025-09-07T07:44:00.7216407Z inflating: build/bin/c10_ssize_test 2025-09-07T07:44:00.7268445Z inflating: build/bin/c10_string_util_test 2025-09-07T07:44:00.7315430Z inflating: build/bin/c10_string_view_test 2025-09-07T07:44:00.7363051Z inflating: build/bin/c10_tempfile_test 2025-09-07T07:44:00.7416470Z inflating: build/bin/c10_typeid_test 2025-09-07T07:44:00.7458411Z inflating: build/bin/c10_intrusive_ptr_benchmark 2025-09-07T07:44:00.7966685Z inflating: build/bin/vec_test_all_types_DEFAULT 2025-09-07T07:44:00.8491835Z inflating: build/bin/vec_test_all_types_AVX512 2025-09-07T07:44:00.9021363Z inflating: build/bin/vec_test_all_types_AVX2 2025-09-07T07:44:00.9070522Z inflating: build/bin/static_runtime_bench 2025-09-07T07:44:00.9293787Z inflating: build/bin/static_runtime_test 2025-09-07T07:44:00.9361088Z inflating: build/bin/Dict_test 2025-09-07T07:44:00.9410723Z inflating: build/bin/Dimname_test 2025-09-07T07:44:00.9471710Z inflating: build/bin/MaybeOwned_test 2025-09-07T07:44:00.9525396Z inflating: build/bin/NamedTensor_test 2025-09-07T07:44:00.9580553Z inflating: build/bin/apply_utils_test 2025-09-07T07:44:00.9636088Z inflating: build/bin/atest 2025-09-07T07:44:00.9696115Z inflating: build/bin/basic 2025-09-07T07:44:00.9748683Z inflating: build/bin/broadcast_test 2025-09-07T07:44:00.9797220Z inflating: build/bin/cpu_allocator_test 2025-09-07T07:44:00.9852042Z inflating: build/bin/cpu_generator_test 2025-09-07T07:44:00.9902612Z inflating: build/bin/cpu_profiling_allocator_test 2025-09-07T07:44:00.9987069Z inflating: build/bin/cpu_rng_test 2025-09-07T07:44:01.0035344Z inflating: build/bin/dlconvertor_test 2025-09-07T07:44:01.0089573Z inflating: build/bin/extension_backend_test 2025-09-07T07:44:01.0141998Z inflating: build/bin/half_test 2025-09-07T07:44:01.0230750Z inflating: build/bin/ivalue_test 2025-09-07T07:44:01.0278754Z inflating: build/bin/lazy_tensor_test 2025-09-07T07:44:01.0329051Z inflating: build/bin/math_kernel_test 2025-09-07T07:44:01.0380175Z inflating: build/bin/memory_format_test 2025-09-07T07:44:01.0431025Z inflating: build/bin/memory_overlapping_test 2025-09-07T07:44:01.0481577Z inflating: build/bin/mobile_memory_cleanup 2025-09-07T07:44:01.0534642Z inflating: build/bin/native_test 2025-09-07T07:44:01.0583103Z inflating: build/bin/operator_name_test 2025-09-07T07:44:01.0631145Z inflating: build/bin/operators_test 2025-09-07T07:44:01.0680760Z inflating: build/bin/packedtensoraccessor_test 2025-09-07T07:44:01.0743288Z inflating: build/bin/pow_test 2025-09-07T07:44:01.0797618Z inflating: build/bin/quantized_test 2025-09-07T07:44:01.0845140Z inflating: build/bin/reduce_ops_test 2025-09-07T07:44:01.0893680Z inflating: build/bin/reportMemoryUsage_test 2025-09-07T07:44:01.0947138Z inflating: build/bin/scalar_tensor_test 2025-09-07T07:44:01.1002518Z inflating: build/bin/scalar_test 2025-09-07T07:44:01.1051267Z inflating: build/bin/StorageUtils_test 2025-09-07T07:44:01.1100934Z inflating: build/bin/stride_properties_test 2025-09-07T07:44:01.1173399Z inflating: build/bin/tensor_iterator_test 2025-09-07T07:44:01.1225140Z inflating: build/bin/test_parallel 2025-09-07T07:44:01.1273023Z inflating: build/bin/thread_init_test 2025-09-07T07:44:01.1325038Z inflating: build/bin/type_ptr_test 2025-09-07T07:44:01.1380857Z inflating: build/bin/type_test 2025-09-07T07:44:01.1430567Z inflating: build/bin/undefined_tensor_test 2025-09-07T07:44:01.1477888Z inflating: build/bin/verify_api_visibility 2025-09-07T07:44:01.1543276Z inflating: build/bin/legacy_vmap_test 2025-09-07T07:44:01.1591676Z inflating: build/bin/weakref_test 2025-09-07T07:44:01.1640725Z inflating: build/bin/wrapdim_test 2025-09-07T07:44:01.1689586Z inflating: build/bin/xla_tensor_test 2025-09-07T07:44:01.1745522Z inflating: build/bin/IListRef_test 2025-09-07T07:44:01.1841670Z inflating: build/bin/List_test 2025-09-07T07:44:01.1903424Z inflating: build/bin/KernelFunction_test 2025-09-07T07:44:01.2011919Z inflating: build/bin/kernel_function_legacy_test 2025-09-07T07:44:01.2098829Z inflating: build/bin/kernel_function_test 2025-09-07T07:44:01.2211883Z inflating: build/bin/kernel_lambda_legacy_test 2025-09-07T07:44:01.2303976Z inflating: build/bin/kernel_lambda_test 2025-09-07T07:44:01.2361060Z inflating: build/bin/kernel_stackbased_test 2025-09-07T07:44:01.2448015Z inflating: build/bin/make_boxed_from_unboxed_functor_test 2025-09-07T07:44:01.2496568Z inflating: build/bin/CppSignature_test 2025-09-07T07:44:01.2548862Z inflating: build/bin/backend_fallback_test 2025-09-07T07:44:01.2821670Z inflating: build/bin/op_registration_test 2025-09-07T07:44:01.2866507Z inflating: build/bin/op_allowlist_test 2025-09-07T07:44:01.2928758Z inflating: build/bin/inline_container_test 2025-09-07T07:44:01.3893793Z inflating: build/bin/test_jit 2025-09-07T07:44:01.3942128Z inflating: build/bin/FileStoreTest 2025-09-07T07:44:01.3992058Z inflating: build/bin/BackoffTest 2025-09-07T07:44:01.4045724Z inflating: build/bin/TCPStoreTest 2025-09-07T07:44:01.4378861Z inflating: build/bin/test_nativert 2025-09-07T07:44:01.4427890Z inflating: build/bin/HashStoreTest 2025-09-07T07:44:01.4430295Z inflating: build/bin/example_allreduce 2025-09-07T07:44:01.4492377Z inflating: build/bin/ProcessGroupGlooTest 2025-09-07T07:44:01.4544164Z inflating: build/bin/test_dist_autograd 2025-09-07T07:44:01.4607722Z inflating: build/bin/test_cpp_rpc 2025-09-07T07:44:01.5597117Z inflating: build/bin/test_api 2025-09-07T07:44:01.5597500Z inflating: build/bin/parallel_benchmark 2025-09-07T07:44:01.5896635Z inflating: build/bin/test_lazy 2025-09-07T07:44:01.5899275Z inflating: build/bin/torch_shm_manager 2025-09-07T07:44:01.5899550Z creating: .additional_ci_files/ 2025-09-07T07:44:01.5975417Z inflating: .additional_ci_files/test-times.json 2025-09-07T07:44:01.6267114Z inflating: .additional_ci_files/test-class-times.json 2025-09-07T07:44:01.6324093Z ##[group]Run rm artifacts.zip 2025-09-07T07:44:01.6324304Z rm artifacts.zip 2025-09-07T07:44:01.6328717Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:44:01.6328940Z env: 2025-09-07T07:44:01.6329092Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:44:01.6329266Z ##[endgroup] 2025-09-07T07:44:01.6658052Z ##[group]Run df -H 2025-09-07T07:44:01.6658226Z df -H 2025-09-07T07:44:01.6661982Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:44:01.6662204Z env: 2025-09-07T07:44:01.6662364Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:44:01.6662538Z ##[endgroup] 2025-09-07T07:44:01.6694509Z Filesystem Size Used Avail Use% Mounted on 2025-09-07T07:44:01.6694791Z devtmpfs 4.2M 0 4.2M 0% /dev 2025-09-07T07:44:01.6695005Z tmpfs 102G 0 102G 0% /dev/shm 2025-09-07T07:44:01.6695220Z tmpfs 41G 11M 41G 1% /run 2025-09-07T07:44:01.6695431Z /dev/nvme0n1p1 215G 72G 144G 34% / 2025-09-07T07:44:01.6695639Z tmpfs 102G 13k 102G 1% /tmp 2025-09-07T07:44:01.6695937Z /dev/nvme0n1p128 11M 1.4M 9.2M 13% /boot/efi 2025-09-07T07:44:01.6716691Z Prepare all required actions 2025-09-07T07:44:01.6717298Z Getting action download info 2025-09-07T07:44:01.8152163Z ##[group]Run ./.github/actions/download-td-artifacts 2025-09-07T07:44:01.8152534Z with: 2025-09-07T07:44:01.8152724Z env: 2025-09-07T07:44:01.8175199Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:44:01.8175402Z ##[endgroup] 2025-09-07T07:44:01.8209551Z ##[group]Run seemethere/download-artifact-s3@v4 2025-09-07T07:44:01.8209765Z with: 2025-09-07T07:44:01.8209899Z name: td_results 2025-09-07T07:44:01.8210060Z s3-bucket: gha-artifacts 2025-09-07T07:44:01.8210233Z region: us-east-1 2025-09-07T07:44:01.8210382Z env: 2025-09-07T07:44:01.8210517Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:44:01.8210678Z ##[endgroup] 2025-09-07T07:44:02.1484771Z (node:57962) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-09-07T07:44:02.1485057Z 2025-09-07T07:44:02.1485237Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-09-07T07:44:02.1485567Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-09-07T07:44:02.2187774Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-09-07T07:44:02.2188129Z Found 0 objects with prefix pytorch/pytorch/17525285611/td_results/ 2025-09-07T07:44:02.2191020Z Artifact download has finished successfully 2025-09-07T07:44:02.2380327Z ##[group]Run mkdir -p .additional_ci_files 2025-09-07T07:44:02.2380571Z mkdir -p .additional_ci_files 2025-09-07T07:44:02.2380824Z mv td_results.json .additional_ci_files/td_results.json || true 2025-09-07T07:44:02.2385484Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:44:02.2385713Z env: 2025-09-07T07:44:02.2385863Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:44:02.2386037Z ##[endgroup] 2025-09-07T07:44:02.2424438Z mv: cannot stat 'td_results.json': No such file or directory 2025-09-07T07:44:02.2481150Z ##[group]Run .github/scripts/parse_ref.py 2025-09-07T07:44:02.2481385Z .github/scripts/parse_ref.py 2025-09-07T07:44:02.2485265Z shell: /usr/bin/bash -e {0} 2025-09-07T07:44:02.2485440Z env: 2025-09-07T07:44:02.2485583Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:44:02.2485745Z ##[endgroup] 2025-09-07T07:44:02.3081242Z Setting output branch=main 2025-09-07T07:44:02.3160736Z Prepare all required actions 2025-09-07T07:44:02.3161018Z Getting action download info 2025-09-07T07:44:02.4311981Z ##[group]Run ./.github/actions/filter-test-configs 2025-09-07T07:44:02.4312192Z with: 2025-09-07T07:44:02.4312504Z github-token: *** 2025-09-07T07:44:02.4314602Z test-matrix: {"include": [{"config": "inductor_huggingface_perf_cpu_x86", "shard": 1, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_huggingface_perf_cpu_x86", "shard": 2, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_huggingface_perf_cpu_x86", "shard": 3, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 1, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 2, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 3, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 4, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 5, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 1, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 2, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 3, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 4, "num_shards": 4, "runner": "linux.24xl.spr-metal"}]} 2025-09-07T07:44:02.4316968Z job-name: inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal) 2025-09-07T07:44:02.4317299Z env: 2025-09-07T07:44:02.4317445Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:44:02.4317616Z ##[endgroup] 2025-09-07T07:44:02.4376826Z ##[group]Run nick-fields/retry@v3.0.0 2025-09-07T07:44:02.4377026Z with: 2025-09-07T07:44:02.4377187Z shell: bash 2025-09-07T07:44:02.4377345Z timeout_minutes: 10 2025-09-07T07:44:02.4377511Z max_attempts: 5 2025-09-07T07:44:02.4377669Z retry_wait_seconds: 30 2025-09-07T07:44:02.4378109Z command: set -eux # PyYAML 6.0 doesn't work with MacOS x86 anymore # This must run on Python-3.7 (AmazonLinux2) so can't use request=3.32.2 python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-09-07T07:44:02.4378560Z polling_interval_seconds: 1 2025-09-07T07:44:02.4378741Z warning_on_retry: true 2025-09-07T07:44:02.4378912Z continue_on_error: false 2025-09-07T07:44:02.4379083Z env: 2025-09-07T07:44:02.4379219Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:44:02.4379515Z GITHUB_TOKEN: *** 2025-09-07T07:44:02.4379670Z ##[endgroup] 2025-09-07T07:44:02.5823026Z + python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-09-07T07:44:02.7429391Z Defaulting to user installation because normal site-packages is not writeable 2025-09-07T07:44:02.8387027Z Collecting requests==2.27.1 2025-09-07T07:44:02.8508557Z Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB) 2025-09-07T07:44:02.9918572Z Collecting pyyaml==6.0.2 2025-09-07T07:44:02.9947682Z Downloading PyYAML-6.0.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (737 kB) 2025-09-07T07:44:03.1437773Z Collecting certifi>=2017.4.17 2025-09-07T07:44:03.1465635Z Downloading certifi-2025.8.3-py3-none-any.whl (161 kB) 2025-09-07T07:44:03.4668174Z Collecting charset-normalizer~=2.0.0 2025-09-07T07:44:03.4696799Z Downloading charset_normalizer-2.0.12-py3-none-any.whl (39 kB) 2025-09-07T07:44:03.5257540Z Requirement already satisfied: urllib3<1.27,>=1.21.1 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (1.25.10) 2025-09-07T07:44:03.5262131Z Requirement already satisfied: idna<4,>=2.5 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (2.10) 2025-09-07T07:44:03.5812032Z Installing collected packages: charset-normalizer, certifi, requests, pyyaml 2025-09-07T07:44:04.0517711Z Successfully installed certifi-2025.8.3 charset-normalizer-2.0.12 pyyaml-6.0.2 requests-2.27.1 2025-09-07T07:44:04.4941135Z Command completed after 1 attempt(s). 2025-09-07T07:44:04.5017106Z ##[group]Run set -x 2025-09-07T07:44:04.5017297Z set -x 2025-09-07T07:44:04.5017449Z  2025-09-07T07:44:04.5017683Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-09-07T07:44:04.5017974Z # in runner workspace 2025-09-07T07:44:04.5018219Z python3 "${GITHUB_ACTION_PATH}/../../scripts/parse_ref.py" 2025-09-07T07:44:04.5022924Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:44:04.5023157Z env: 2025-09-07T07:44:04.5023302Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:44:04.5023480Z ##[endgroup] 2025-09-07T07:44:04.5043075Z + python3 /home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/filter-test-configs/../../scripts/parse_ref.py 2025-09-07T07:44:04.5173357Z Setting output branch=main 2025-09-07T07:44:04.5238752Z ##[group]Run echo "Workflow: ${GITHUB_WORKFLOW}" 2025-09-07T07:44:04.5239039Z echo "Workflow: ${GITHUB_WORKFLOW}" 2025-09-07T07:44:04.5239253Z echo "Job name: ${JOB_NAME}" 2025-09-07T07:44:04.5239443Z  2025-09-07T07:44:04.5239690Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-09-07T07:44:04.5239978Z # in runner workspace 2025-09-07T07:44:04.5240262Z python3 "${GITHUB_ACTION_PATH}/../../scripts/filter_test_configs.py" \ 2025-09-07T07:44:04.5240554Z  --workflow "${GITHUB_WORKFLOW}" \ 2025-09-07T07:44:04.5240759Z  --job-name "${JOB_NAME}" \ 2025-09-07T07:44:04.5242985Z  --test-matrix "{"include": [{"config": "inductor_huggingface_perf_cpu_x86", "shard": 1, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_huggingface_perf_cpu_x86", "shard": 2, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_huggingface_perf_cpu_x86", "shard": 3, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 1, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 2, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 3, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 4, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 5, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 1, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 2, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 3, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 4, "num_shards": 4, "runner": "linux.24xl.spr-metal"}]}" \ 2025-09-07T07:44:04.5245449Z  --selected-test-configs "" \ 2025-09-07T07:44:04.5245692Z  --pr-number "${PR_NUMBER}" \ 2025-09-07T07:44:04.5245913Z  --tag "${TAG}" \ 2025-09-07T07:44:04.5246118Z  --event-name "${EVENT_NAME}" \ 2025-09-07T07:44:04.5246346Z  --schedule "${SCHEDULE}" \ 2025-09-07T07:44:04.5246576Z  --branch "${HEAD_BRANCH}" 2025-09-07T07:44:04.5251054Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:44:04.5251298Z env: 2025-09-07T07:44:04.5251461Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:44:04.5251852Z GITHUB_TOKEN: *** 2025-09-07T07:44:04.5252219Z JOB_NAME: inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal) 2025-09-07T07:44:04.5252618Z PR_NUMBER: 2025-09-07T07:44:04.5252774Z TAG: 2025-09-07T07:44:04.5252927Z EVENT_NAME: schedule 2025-09-07T07:44:04.5253108Z SCHEDULE: 0 7 * * * 2025-09-07T07:44:04.5253279Z HEAD_BRANCH: main 2025-09-07T07:44:04.5253443Z ##[endgroup] 2025-09-07T07:44:04.5270343Z Workflow: inductor-perf-nightly-x86 2025-09-07T07:44:04.5270825Z Job name: inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal) 2025-09-07T07:44:04.6743767Z Setting output keep-going=True 2025-09-07T07:44:04.6744136Z Setting output ci-verbose-test-logs=False 2025-09-07T07:44:04.6744409Z Setting output ci-test-showlocals=False 2025-09-07T07:44:04.6744649Z Setting output ci-no-test-timeout=False 2025-09-07T07:44:04.6744869Z Setting output ci-no-td=False 2025-09-07T07:44:04.6745077Z Setting output ci-td-distributed=False 2025-09-07T07:44:04.6745290Z Setting output is-unstable=False 2025-09-07T07:44:04.6745477Z Setting output reenabled-issues= 2025-09-07T07:44:04.6747667Z Setting output test-matrix={"include": [{"config": "inductor_huggingface_perf_cpu_x86", "shard": 1, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_huggingface_perf_cpu_x86", "shard": 2, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_huggingface_perf_cpu_x86", "shard": 3, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 1, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 2, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 3, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 4, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 5, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 1, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 2, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 3, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 4, "num_shards": 4, "runner": "linux.24xl.spr-metal"}]} 2025-09-07T07:44:04.6750014Z Setting output is-test-matrix-empty=False 2025-09-07T07:44:04.6874335Z ##[group]Run echo "Filtered matrix:" 2025-09-07T07:44:04.6874563Z echo "Filtered matrix:" 2025-09-07T07:44:04.6876646Z echo "{"include": [{"config": "inductor_huggingface_perf_cpu_x86", "shard": 1, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_huggingface_perf_cpu_x86", "shard": 2, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_huggingface_perf_cpu_x86", "shard": 3, "num_shards": 3, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 1, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 2, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 3, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 4, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_timm_perf_cpu_x86", "shard": 5, "num_shards": 5, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 1, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 2, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 3, "num_shards": 4, "runner": "linux.24xl.spr-metal"}, {"config": "inductor_torchbench_perf_cpu_x86", "shard": 4, "num_shards": 4, "runner": "linux.24xl.spr-metal"}]}" 2025-09-07T07:44:04.6879004Z  2025-09-07T07:44:04.6879155Z echo 2025-09-07T07:44:04.6879342Z echo "Is the current job unstable? False" 2025-09-07T07:44:04.6879549Z  2025-09-07T07:44:04.6879681Z echo 2025-09-07T07:44:04.6879854Z echo "Is keep-going label set? True" 2025-09-07T07:44:04.6880059Z  2025-09-07T07:44:04.6880187Z echo 2025-09-07T07:44:04.6880387Z echo "Reenabled issues? " 2025-09-07T07:44:04.6883859Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:44:04.6884088Z env: 2025-09-07T07:44:04.6884235Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:44:04.6884398Z ##[endgroup] 2025-09-07T07:44:04.6900673Z Filtered matrix: 2025-09-07T07:44:04.6903218Z {include: [{config: inductor_huggingface_perf_cpu_x86, shard: 1, num_shards: 3, runner: linux.24xl.spr-metal}, {config: inductor_huggingface_perf_cpu_x86, shard: 2, num_shards: 3, runner: linux.24xl.spr-metal}, {config: inductor_huggingface_perf_cpu_x86, shard: 3, num_shards: 3, runner: linux.24xl.spr-metal}, {config: inductor_timm_perf_cpu_x86, shard: 1, num_shards: 5, runner: linux.24xl.spr-metal}, {config: inductor_timm_perf_cpu_x86, shard: 2, num_shards: 5, runner: linux.24xl.spr-metal}, {config: inductor_timm_perf_cpu_x86, shard: 3, num_shards: 5, runner: linux.24xl.spr-metal}, {config: inductor_timm_perf_cpu_x86, shard: 4, num_shards: 5, runner: linux.24xl.spr-metal}, {config: inductor_timm_perf_cpu_x86, shard: 5, num_shards: 5, runner: linux.24xl.spr-metal}, {config: inductor_torchbench_perf_cpu_x86, shard: 1, num_shards: 4, runner: linux.24xl.spr-metal}, {config: inductor_torchbench_perf_cpu_x86, shard: 2, num_shards: 4, runner: linux.24xl.spr-metal}, {config: inductor_torchbench_perf_cpu_x86, shard: 3, num_shards: 4, runner: linux.24xl.spr-metal}, {config: inductor_torchbench_perf_cpu_x86, shard: 4, num_shards: 4, runner: linux.24xl.spr-metal}]} 2025-09-07T07:44:04.6905399Z 2025-09-07T07:44:04.6905489Z Is the current job unstable? False 2025-09-07T07:44:04.6905626Z 2025-09-07T07:44:04.6905698Z Is keep-going label set? True 2025-09-07T07:44:04.6905822Z 2025-09-07T07:44:04.6905884Z Reenabled issues? 2025-09-07T07:44:04.7081970Z ##[group]Run echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-09-07T07:44:04.7082286Z echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-09-07T07:44:04.7086171Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:44:04.7086411Z env: 2025-09-07T07:44:04.7086560Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:44:04.7086727Z JOB_TIMEOUT: 720 2025-09-07T07:44:04.7086875Z ##[endgroup] 2025-09-07T07:44:04.7261429Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:44:04.7261758Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:44:04.7262023Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:44:04.7265795Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:44:04.7266021Z env: 2025-09-07T07:44:04.7266169Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:44:04.7266336Z ##[endgroup] 2025-09-07T07:44:04.7437678Z ##[group]Run set -x 2025-09-07T07:44:04.7437928Z set -x 2025-09-07T07:44:04.7438078Z  2025-09-07T07:44:04.7438258Z if [[ $TEST_CONFIG == 'multigpu' ]]; then 2025-09-07T07:44:04.7438512Z  TEST_COMMAND=.ci/pytorch/multigpu-test.sh 2025-09-07T07:44:04.7438837Z elif [[ $BUILD_ENVIRONMENT == *onnx* ]]; then 2025-09-07T07:44:04.7439061Z  TEST_COMMAND=.ci/onnx/test.sh 2025-09-07T07:44:04.7439258Z else 2025-09-07T07:44:04.7439428Z  TEST_COMMAND=.ci/pytorch/test.sh 2025-09-07T07:44:04.7439624Z fi 2025-09-07T07:44:04.7439756Z  2025-09-07T07:44:04.7439935Z # Leaving 1GB for the runner and other things 2025-09-07T07:44:04.7440289Z TOTAL_AVAILABLE_MEMORY_IN_GB=$(awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo) 2025-09-07T07:44:04.7440814Z # https://docs.docker.com/engine/containers/resource_constraints/#--memory-swap-details, the 3GB swap 2025-09-07T07:44:04.7441221Z # comes from https://github.com/pytorch/test-infra/pull/6058 2025-09-07T07:44:04.7441542Z TOTAL_MEMORY_WITH_SWAP=$(("${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}" + 3)) 2025-09-07T07:44:04.7441789Z  2025-09-07T07:44:04.7441967Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-09-07T07:44:04.7442179Z  SHM_OPTS= 2025-09-07T07:44:04.7442347Z  JENKINS_USER= 2025-09-07T07:44:04.7442573Z  # ensure that docker container cleanly exits in 12 hours 2025-09-07T07:44:04.7442858Z  # if for some reason cleanup action doesn't stop container 2025-09-07T07:44:04.7443098Z  # when job is cancelled 2025-09-07T07:44:04.7443291Z  DOCKER_SHELL_CMD="sleep 12h" 2025-09-07T07:44:04.7443478Z else 2025-09-07T07:44:04.7443652Z  SHM_OPTS="--shm-size=${SHM_SIZE}" 2025-09-07T07:44:04.7443865Z  JENKINS_USER="--user jenkins" 2025-09-07T07:44:04.7444062Z  DOCKER_SHELL_CMD= 2025-09-07T07:44:04.7444237Z fi 2025-09-07T07:44:04.7444378Z  2025-09-07T07:44:04.7444589Z # detached container should get cleaned up by teardown_ec2_linux 2025-09-07T07:44:04.7444901Z # TODO: Stop building test binaries as part of the build phase 2025-09-07T07:44:04.7445260Z # Used for GPU_FLAG, SHM_OPTS, JENKINS_USER and DOCKER_SHELL_CMD since that doesn't play nice 2025-09-07T07:44:04.7445577Z # shellcheck disable=SC2086,SC2090 2025-09-07T07:44:04.7445788Z container_name=$(docker run \ 2025-09-07T07:44:04.7445990Z  ${GPU_FLAG:-} \ 2025-09-07T07:44:04.7446184Z  ${SCCACHE_SERVER_PORT_DOCKER_FLAG:-} \ 2025-09-07T07:44:04.7446401Z  -e BUILD_ENVIRONMENT \ 2025-09-07T07:44:04.7446597Z  -e PR_NUMBER \ 2025-09-07T07:44:04.7446777Z  -e GITHUB_ACTIONS \ 2025-09-07T07:44:04.7446963Z  -e GITHUB_REPOSITORY \ 2025-09-07T07:44:04.7447162Z  -e GITHUB_WORKFLOW \ 2025-09-07T07:44:04.7447351Z  -e GITHUB_JOB \ 2025-09-07T07:44:04.7447526Z  -e GITHUB_RUN_ID \ 2025-09-07T07:44:04.7447702Z  -e GITHUB_RUN_NUMBER \ 2025-09-07T07:44:04.7447894Z  -e GITHUB_RUN_ATTEMPT \ 2025-09-07T07:44:04.7448088Z  -e JOB_ID \ 2025-09-07T07:44:04.7448255Z  -e JOB_NAME \ 2025-09-07T07:44:04.7448419Z  -e BASE_SHA \ 2025-09-07T07:44:04.7448584Z  -e BRANCH \ 2025-09-07T07:44:04.7448744Z  -e SHA1 \ 2025-09-07T07:44:04.7448906Z  -e AWS_DEFAULT_REGION \ 2025-09-07T07:44:04.7449086Z  -e IN_WHEEL_TEST \ 2025-09-07T07:44:04.7449278Z  -e SHARD_NUMBER \ 2025-09-07T07:44:04.7449457Z  -e TEST_CONFIG \ 2025-09-07T07:44:04.7449641Z  -e NUM_TEST_SHARDS \ 2025-09-07T07:44:04.7449831Z  -e REENABLED_ISSUES \ 2025-09-07T07:44:04.7450018Z  -e CONTINUE_THROUGH_ERROR \ 2025-09-07T07:44:04.7450284Z  -e VERBOSE_TEST_LOGS \ 2025-09-07T07:44:04.7450474Z  -e TEST_SHOWLOCALS \ 2025-09-07T07:44:04.7450654Z  -e NO_TEST_TIMEOUT \ 2025-09-07T07:44:04.7450820Z  -e NO_TD \ 2025-09-07T07:44:04.7450990Z  -e TD_DISTRIBUTED \ 2025-09-07T07:44:04.7451203Z  -e PR_LABELS \ 2025-09-07T07:44:04.7451398Z  -e MAX_JOBS="$(nproc --ignore=2)" \ 2025-09-07T07:44:04.7451598Z  -e SCCACHE_BUCKET \ 2025-09-07T07:44:04.7451782Z  -e SCCACHE_REGION \ 2025-09-07T07:44:04.7451960Z  -e XLA_CUDA \ 2025-09-07T07:44:04.7452152Z  -e XLA_CLANG_CACHE_S3_BUCKET_NAME \ 2025-09-07T07:44:04.7452367Z  -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK \ 2025-09-07T07:44:04.7452599Z  -e PYTORCH_TEST_RERUN_DISABLED_TESTS \ 2025-09-07T07:44:04.7452830Z  -e SKIP_SCCACHE_INITIALIZATION=1 \ 2025-09-07T07:44:04.7453045Z  -e HUGGING_FACE_HUB_TOKEN \ 2025-09-07T07:44:04.7453254Z  -e VLLM_TEST_HUGGING_FACE_TOKEN \ 2025-09-07T07:44:04.7453460Z  -e SCRIBE_GRAPHQL_ACCESS_TOKEN \ 2025-09-07T07:44:04.7453659Z  -e DASHBOARD_TAG \ 2025-09-07T07:44:04.7453845Z  -e ARTIFACTS_FILE_SUFFIX \ 2025-09-07T07:44:04.7454075Z  --memory="${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}g" \ 2025-09-07T07:44:04.7454330Z  --memory-swap="${TOTAL_MEMORY_WITH_SWAP}g" \ 2025-09-07T07:44:04.7454583Z  --env-file="/tmp/github_env_${GITHUB_RUN_ID}" \ 2025-09-07T07:44:04.7454826Z  --security-opt seccomp=unconfined \ 2025-09-07T07:44:04.7455031Z  --cap-add=SYS_PTRACE \ 2025-09-07T07:44:04.7455211Z  --ipc=host \ 2025-09-07T07:44:04.7455378Z  ${SHM_OPTS} \ 2025-09-07T07:44:04.7455542Z  --tty \ 2025-09-07T07:44:04.7455700Z  --detach \ 2025-09-07T07:44:04.7455869Z  --name="${container_name}" \ 2025-09-07T07:44:04.7456069Z  ${JENKINS_USER} \ 2025-09-07T07:44:04.7456291Z  -v "${GITHUB_WORKSPACE}:/var/lib/jenkins/workspace" \ 2025-09-07T07:44:04.7456537Z  -w /var/lib/jenkins/workspace \ 2025-09-07T07:44:04.7456734Z  "${DOCKER_IMAGE}" \ 2025-09-07T07:44:04.7456904Z  ${DOCKER_SHELL_CMD} 2025-09-07T07:44:04.7457075Z ) 2025-09-07T07:44:04.7457272Z # Propagate download.pytorch.org IP to container 2025-09-07T07:44:04.7457664Z grep download.pytorch.org /etc/hosts | docker exec -i "${container_name}" sudo bash -c "/bin/cat >> /etc/hosts" 2025-09-07T07:44:04.7458062Z echo "DOCKER_CONTAINER_ID=${container_name}" >> "${GITHUB_ENV}" 2025-09-07T07:44:04.7458308Z  2025-09-07T07:44:04.7458485Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-09-07T07:44:04.7458821Z  docker exec -t "${container_name}" sh -c "python3 -m pip install -r .ci/docker/requirements-ci.txt" 2025-09-07T07:44:04.7459118Z fi 2025-09-07T07:44:04.7459248Z  2025-09-07T07:44:04.7459539Z docker exec -t "${container_name}" sh -c "python3 -m pip install $(echo dist/*.whl)[opt-einsum] && ${TEST_COMMAND}" 2025-09-07T07:44:04.7463198Z shell: /usr/bin/bash -e {0} 2025-09-07T07:44:04.7463372Z env: 2025-09-07T07:44:04.7463515Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:44:04.7463743Z BUILD_ENVIRONMENT: linux-jammy-py3.9-gcc11-build 2025-09-07T07:44:04.7463966Z PR_NUMBER: 2025-09-07T07:44:04.7464133Z GITHUB_REPOSITORY: pytorch/pytorch 2025-09-07T07:44:04.7464349Z GITHUB_WORKFLOW: inductor-perf-nightly-x86 2025-09-07T07:44:04.7464560Z GITHUB_JOB: test 2025-09-07T07:44:04.7464718Z GITHUB_RUN_ID: 17525285611 2025-09-07T07:44:04.7464893Z GITHUB_RUN_NUMBER: 525 2025-09-07T07:44:04.7465054Z GITHUB_RUN_ATTEMPT: 1 2025-09-07T07:44:04.7465219Z JOB_ID: 49775585800 2025-09-07T07:44:04.7465549Z JOB_NAME: inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal) 2025-09-07T07:44:04.7465893Z BRANCH: main 2025-09-07T07:44:04.7466137Z SHA1: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:44:04.7466370Z BASE_SHA: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:44:04.7466614Z TEST_CONFIG: inductor_torchbench_perf_cpu_x86 2025-09-07T07:44:04.7466820Z SHARD_NUMBER: 2 2025-09-07T07:44:04.7466978Z NUM_TEST_SHARDS: 4 2025-09-07T07:44:04.7467170Z REENABLED_ISSUES: 2025-09-07T07:44:04.7467341Z CONTINUE_THROUGH_ERROR: True 2025-09-07T07:44:04.7467530Z VERBOSE_TEST_LOGS: False 2025-09-07T07:44:04.7467709Z TEST_SHOWLOCALS: False 2025-09-07T07:44:04.7467877Z NO_TEST_TIMEOUT: False 2025-09-07T07:44:04.7468042Z NO_TD: False 2025-09-07T07:44:04.7468200Z TD_DISTRIBUTED: False 2025-09-07T07:44:04.7468479Z SCCACHE_BUCKET: ossci-compiler-cache-circleci-v2 2025-09-07T07:44:04.7468700Z SCCACHE_REGION: us-east-1 2025-09-07T07:44:04.7468878Z SHM_SIZE: 1g 2025-09-07T07:44:04.7469364Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:44:04.7469863Z XLA_CUDA: 2025-09-07T07:44:04.7470088Z XLA_CLANG_CACHE_S3_BUCKET_NAME: ossci-compiler-clang-cache-circleci-xla 2025-09-07T07:44:04.7470370Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: 0 2025-09-07T07:44:04.7470583Z PYTORCH_TEST_RERUN_DISABLED_TESTS: 0 2025-09-07T07:44:04.7471032Z DASHBOARD_TAG: training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true 2025-09-07T07:44:04.7471625Z VLLM_TEST_HUGGING_FACE_TOKEN: *** 2025-09-07T07:44:04.7471888Z HUGGING_FACE_HUB_TOKEN: *** 2025-09-07T07:44:04.7472151Z SCRIBE_GRAPHQL_ACCESS_TOKEN: *** 2025-09-07T07:44:04.7472468Z ARTIFACTS_FILE_SUFFIX: test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800 2025-09-07T07:44:04.7472792Z ##[endgroup] 2025-09-07T07:44:04.7488072Z + [[ inductor_torchbench_perf_cpu_x86 == \m\u\l\t\i\g\p\u ]] 2025-09-07T07:44:04.7488339Z + [[ linux-jammy-py3.9-gcc11-build == *onnx* ]] 2025-09-07T07:44:04.7488563Z + TEST_COMMAND=.ci/pytorch/test.sh 2025-09-07T07:44:04.7492453Z ++ awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo 2025-09-07T07:44:04.7506395Z + TOTAL_AVAILABLE_MEMORY_IN_GB='187.488 ' 2025-09-07T07:44:04.7506623Z + TOTAL_MEMORY_WITH_SWAP=190 2025-09-07T07:44:04.7506847Z + [[ linux-jammy-py3.9-gcc11-build == *\s\3\9\0\x* ]] 2025-09-07T07:44:04.7507090Z + SHM_OPTS=--shm-size=1g 2025-09-07T07:44:04.7507267Z + JENKINS_USER='--user jenkins' 2025-09-07T07:44:04.7507449Z + DOCKER_SHELL_CMD= 2025-09-07T07:44:04.7514637Z +++ nproc --ignore=2 2025-09-07T07:44:04.7839743Z ++ docker run -e BUILD_ENVIRONMENT -e PR_NUMBER -e GITHUB_ACTIONS -e GITHUB_REPOSITORY -e GITHUB_WORKFLOW -e GITHUB_JOB -e GITHUB_RUN_ID -e GITHUB_RUN_NUMBER -e GITHUB_RUN_ATTEMPT -e JOB_ID -e JOB_NAME -e BASE_SHA -e BRANCH -e SHA1 -e AWS_DEFAULT_REGION -e IN_WHEEL_TEST -e SHARD_NUMBER -e TEST_CONFIG -e NUM_TEST_SHARDS -e REENABLED_ISSUES -e CONTINUE_THROUGH_ERROR -e VERBOSE_TEST_LOGS -e TEST_SHOWLOCALS -e NO_TEST_TIMEOUT -e NO_TD -e TD_DISTRIBUTED -e PR_LABELS -e MAX_JOBS=94 -e SCCACHE_BUCKET -e SCCACHE_REGION -e XLA_CUDA -e XLA_CLANG_CACHE_S3_BUCKET_NAME -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK -e PYTORCH_TEST_RERUN_DISABLED_TESTS -e SKIP_SCCACHE_INITIALIZATION=1 -e HUGGING_FACE_HUB_TOKEN -e VLLM_TEST_HUGGING_FACE_TOKEN -e SCRIBE_GRAPHQL_ACCESS_TOKEN -e DASHBOARD_TAG -e ARTIFACTS_FILE_SUFFIX --memory=187g --memory-swap=190g --env-file=/tmp/github_env_17525285611 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --ipc=host --shm-size=1g --tty --detach --name= --user jenkins -v /home/ec2-user/actions-runner/_work/pytorch/pytorch:/var/lib/jenkins/workspace -w /var/lib/jenkins/workspace 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:46:40.9390909Z + container_name=5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T07:46:40.9393262Z + docker exec -i 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 sudo bash -c '/bin/cat >> /etc/hosts' 2025-09-07T07:46:40.9396266Z + grep download.pytorch.org /etc/hosts 2025-09-07T07:46:41.0786732Z + echo DOCKER_CONTAINER_ID=5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T07:46:41.0787538Z + [[ linux-jammy-py3.9-gcc11-build == *\s\3\9\0\x* ]] 2025-09-07T07:46:41.0790480Z ++ echo dist/torch-2.9.0a0+git93fb23d-cp39-cp39-linux_x86_64.whl 2025-09-07T07:46:41.0792211Z + docker exec -t 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 sh -c 'python3 -m pip install dist/torch-2.9.0a0+git93fb23d-cp39-cp39-linux_x86_64.whl[opt-einsum] && .ci/pytorch/test.sh' 2025-09-07T07:46:41.3733584Z Processing ./dist/torch-2.9.0a0+git93fb23d-cp39-cp39-linux_x86_64.whl (from torch==2.9.0a0+git93fb23d) 2025-09-07T07:46:41.5909202Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (3.19.1) 2025-09-07T07:46:41.5909981Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (4.15.0) 2025-09-07T07:46:41.5912784Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (1.13.3) 2025-09-07T07:46:41.5917907Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (2.8.8) 2025-09-07T07:46:41.5922076Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (3.1.6) 2025-09-07T07:46:41.5927283Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (2025.3.0) 2025-09-07T07:46:41.5943298Z Requirement already satisfied: opt-einsum>=3.3 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (3.3.0) 2025-09-07T07:46:41.6319690Z Requirement already satisfied: numpy>=1.7 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from opt-einsum>=3.3->torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (1.22.4) 2025-09-07T07:46:41.6325835Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from sympy>=1.13.3->torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (1.3.0) 2025-09-07T07:46:41.6354093Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from jinja2->torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (3.0.2) 2025-09-07T07:46:42.3272339Z Installing collected packages: torch 2025-09-07T07:46:48.8464863Z ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts. 2025-09-07T07:46:48.8465482Z dall-e 0.1 requires torchvision, which is not installed. 2025-09-07T07:46:48.8465752Z effdet 0.4.1 requires torchvision, which is not installed. 2025-09-07T07:46:48.8466086Z pytorch-labs-segment-anything-fast 0.2 requires torchao, which is not installed. 2025-09-07T07:46:48.8466527Z pytorch-labs-segment-anything-fast 0.2 requires torchvision>=0.17.0.dev20231026, which is not installed. 2025-09-07T07:46:48.8466977Z timm 1.0.14 requires torchvision, which is not installed. 2025-09-07T07:46:48.8467297Z Successfully installed torch-2.9.0a0+git93fb23d 2025-09-07T07:46:48.9354762Z + export TERM=vt100 2025-09-07T07:46:48.9354983Z + TERM=vt100 2025-09-07T07:46:48.9355152Z ++ dirname .ci/pytorch/test.sh 2025-09-07T07:46:48.9358530Z + source .ci/pytorch/common.sh 2025-09-07T07:46:48.9361709Z +++ dirname .ci/pytorch/common.sh 2025-09-07T07:46:48.9367282Z ++ source .ci/pytorch/common_utils.sh 2025-09-07T07:46:48.9368188Z +++ declare -f -t trap_add 2025-09-07T07:46:48.9371577Z ++ set -ex -o pipefail 2025-09-07T07:46:48.9371801Z ++ [[ linux-jammy-py3.9-gcc11-build == *rocm* ]] 2025-09-07T07:46:48.9372031Z ++ BUILD_TEST_LIBTORCH=0 2025-09-07T07:46:48.9373525Z ++ dirname .ci/pytorch/test.sh 2025-09-07T07:46:48.9377809Z + source .ci/pytorch/common-build.sh 2025-09-07T07:46:48.9379096Z ++ [[ linux-jammy-py3.9-gcc11-build != *win-* ]] 2025-09-07T07:46:48.9386391Z ++++ dirname .ci/pytorch/common-build.sh 2025-09-07T07:46:48.9391949Z +++ cd .ci/pytorch 2025-09-07T07:46:48.9392357Z +++ pwd -P 2025-09-07T07:46:48.9393743Z ++ script_dir=/var/lib/jenkins/workspace/.ci/pytorch 2025-09-07T07:46:48.9394043Z ++ [[ linux-jammy-py3.9-gcc11-build == *-pch* ]] 2025-09-07T07:46:48.9394260Z ++ which sccache 2025-09-07T07:46:48.9408932Z ++ [[ -z ossci-compiler-cache-circleci-v2 ]] 2025-09-07T07:46:48.9409154Z ++ sccache --stop-server 2025-09-07T07:46:48.9430932Z ++ true 2025-09-07T07:46:48.9431106Z ++ rm -f /var/lib/jenkins/sccache_error.log 2025-09-07T07:46:48.9438430Z ++ trap_add sccache_epilogue EXIT 2025-09-07T07:46:48.9438649Z ++ trap_add_cmd=sccache_epilogue 2025-09-07T07:46:48.9438836Z ++ shift 2025-09-07T07:46:48.9438997Z ++ for trap_add_name in "$@" 2025-09-07T07:46:48.9445434Z ++++ trap -p EXIT 2025-09-07T07:46:48.9446644Z +++ eval 'extract_trap_cmd ' 2025-09-07T07:46:48.9446832Z ++++ extract_trap_cmd 2025-09-07T07:46:48.9447005Z ++++ printf '%s\n' '' 2025-09-07T07:46:48.9447184Z +++ printf '%s\n' sccache_epilogue 2025-09-07T07:46:48.9448209Z ++ trap -- ' 2025-09-07T07:46:48.9448371Z sccache_epilogue' EXIT 2025-09-07T07:46:48.9448547Z ++ [[ -n 1 ]] 2025-09-07T07:46:48.9448798Z ++ echo 'Skipping sccache server initialization, setting environment variables' 2025-09-07T07:46:48.9449151Z Skipping sccache server initialization, setting environment variables 2025-09-07T07:46:48.9449412Z ++ export SCCACHE_IDLE_TIMEOUT=0 2025-09-07T07:46:48.9449604Z ++ SCCACHE_IDLE_TIMEOUT=0 2025-09-07T07:46:48.9449829Z ++ export SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-09-07T07:46:48.9450107Z ++ SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-09-07T07:46:48.9450378Z ++ export RUST_LOG=sccache::server=error 2025-09-07T07:46:48.9450585Z ++ RUST_LOG=sccache::server=error 2025-09-07T07:46:48.9450774Z ++ sccache --zero-stats 2025-09-07T07:46:49.0825193Z Statistics zeroed. 2025-09-07T07:46:49.0830241Z ++ which ccache 2025-09-07T07:46:49.0850443Z + [[ linux-jammy-py3.9-gcc11-build != *rocm* ]] 2025-09-07T07:46:49.0850723Z + [[ linux-jammy-py3.9-gcc11-build != *s390x* ]] 2025-09-07T07:46:49.0851000Z + [[ -d /var/lib/jenkins/workspace ]] 2025-09-07T07:46:49.0853240Z ++ stat -c %u /var/lib/jenkins/workspace 2025-09-07T07:46:49.0862284Z + WORKSPACE_ORIGINAL_OWNER_ID=1000 2025-09-07T07:46:49.0862530Z + trap_add cleanup_workspace EXIT 2025-09-07T07:46:49.0862754Z + trap_add_cmd=cleanup_workspace 2025-09-07T07:46:49.0862968Z + shift 2025-09-07T07:46:49.0863150Z + for trap_add_name in "$@" 2025-09-07T07:46:49.0869325Z +++ trap -p EXIT 2025-09-07T07:46:49.0871107Z ++ eval 'extract_trap_cmd trap -- '\'' 2025-09-07T07:46:49.0871363Z sccache_epilogue'\'' EXIT' 2025-09-07T07:46:49.0871572Z +++ extract_trap_cmd trap -- ' 2025-09-07T07:46:49.0871778Z sccache_epilogue' EXIT 2025-09-07T07:46:49.0871960Z +++ printf '%s\n' ' 2025-09-07T07:46:49.0872140Z sccache_epilogue' 2025-09-07T07:46:49.0872340Z ++ printf '%s\n' cleanup_workspace 2025-09-07T07:46:49.0873181Z + trap -- ' 2025-09-07T07:46:49.0873355Z sccache_epilogue 2025-09-07T07:46:49.0873530Z cleanup_workspace' EXIT 2025-09-07T07:46:49.0873767Z + sudo chown -R jenkins /var/lib/jenkins/workspace 2025-09-07T07:46:49.4862446Z + git config --global --add safe.directory /var/lib/jenkins/workspace 2025-09-07T07:46:49.4869450Z + echo 'Environment variables:' 2025-09-07T07:46:49.4869650Z Environment variables: 2025-09-07T07:46:49.4869820Z + env 2025-09-07T07:46:49.4878660Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T07:46:49.4878999Z CONTINUE_THROUGH_ERROR=True 2025-09-07T07:46:49.4879241Z BUILD_ENVIRONMENT=linux-jammy-py3.9-gcc11-build 2025-09-07T07:46:49.4879951Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-09-07T07:46:49.4880164Z HOSTNAME=5b1006e7d104 2025-09-07T07:46:49.4880565Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_0a6d5ec5-039b-4965-a10d-8df077d2d803 2025-09-07T07:46:49.4880933Z GITHUB_ACTION=__run_2 2025-09-07T07:46:49.4881233Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-09-07T07:46:49.4881429Z GITHUB_RUN_NUMBER=525 2025-09-07T07:46:49.4881612Z TEST_CONFIG=inductor_torchbench_perf_cpu_x86 2025-09-07T07:46:49.4881836Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-09-07T07:46:49.4882055Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-09-07T07:46:49.4882262Z SCCACHE_IDLE_TIMEOUT=0 2025-09-07T07:46:49.4882537Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-09-07T07:46:49.4882739Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-09-07T07:46:49.4882940Z GITHUB_REF_TYPE=branch 2025-09-07T07:46:49.4883135Z BASE_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:49.4883342Z XLA_CUDA= 2025-09-07T07:46:49.4883508Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-09-07T07:46:49.4883770Z HUGGING_FACE_HUB_TOKEN=*** 2025-09-07T07:46:49.4887625Z *** 2025-09-07T07:46:49.4887789Z GITHUB_REPOSITORY_ID=65600975 2025-09-07T07:46:49.4887970Z GITHUB_ACTIONS=true 2025-09-07T07:46:49.4888173Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-09-07T07:46:49.4888438Z SHA1=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:49.4888680Z GITHUB_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:49.4889062Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/inductor-perf-test-nightly-x86.yml@refs/heads/main 2025-09-07T07:46:49.4889414Z UCC_HOME=/usr 2025-09-07T07:46:49.4889568Z VERBOSE_TEST_LOGS=False 2025-09-07T07:46:49.4889743Z GITHUB_REF=refs/heads/main 2025-09-07T07:46:49.4889909Z SHARD_NUMBER=2 2025-09-07T07:46:49.4890068Z GITHUB_REF_PROTECTED=true 2025-09-07T07:46:49.4890247Z HOME=/var/lib/jenkins 2025-09-07T07:46:49.4890437Z GITHUB_API_URL=https://api.github.com 2025-09-07T07:46:49.4890644Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-09-07T07:46:49.4890836Z UCX_COMMIT= 2025-09-07T07:46:49.4890980Z USE_SYSTEM_NCCL=1 2025-09-07T07:46:49.4891132Z NUM_TEST_SHARDS=4 2025-09-07T07:46:49.4891276Z UCX_HOME=/usr 2025-09-07T07:46:49.4891617Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_0a6d5ec5-039b-4965-a10d-8df077d2d803 2025-09-07T07:46:49.4892162Z JOB_NAME=inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal) 2025-09-07T07:46:49.4892685Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_0a6d5ec5-039b-4965-a10d-8df077d2d803 2025-09-07T07:46:49.4893137Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-09-07T07:46:49.4893440Z GITHUB_EVENT_NAME=schedule 2025-09-07T07:46:49.4893875Z DASHBOARD_TAG=training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true 2025-09-07T07:46:49.4894314Z GITHUB_RUN_ID=17525285611 2025-09-07T07:46:49.4894491Z INSTALLED_OPENBLAS= 2025-09-07T07:46:49.4894848Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_0a6d5ec5-039b-4965-a10d-8df077d2d803 2025-09-07T07:46:49.4895241Z GITHUB_ACTOR=pytorchmergebot 2025-09-07T07:46:49.4895422Z PR_NUMBER= 2025-09-07T07:46:49.4895567Z DESIRED_CUDA= 2025-09-07T07:46:49.4895712Z GITHUB_RUN_ATTEMPT=1 2025-09-07T07:46:49.4895885Z ANACONDA_PYTHON_VERSION=3.9 2025-09-07T07:46:49.4896099Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-09-07T07:46:49.4896339Z TERM=vt100 2025-09-07T07:46:49.4896491Z INSTALLED_VISION=yes 2025-09-07T07:46:49.4896643Z BRANCH=main 2025-09-07T07:46:49.4896800Z SCCACHE_REGION=us-east-1 2025-09-07T07:46:49.4896981Z OPENSSL_ROOT_DIR=/opt/openssl 2025-09-07T07:46:49.4897164Z CUDA_PATH=/usr/local/cuda 2025-09-07T07:46:49.4897471Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-09-07T07:46:49.4897834Z GITHUB_SERVER_URL=https://github.com 2025-09-07T07:46:49.4898105Z UCC_COMMIT= 2025-09-07T07:46:49.4898256Z REENABLED_ISSUES= 2025-09-07T07:46:49.4898401Z DOCS=yes 2025-09-07T07:46:49.4898544Z SHLVL=1 2025-09-07T07:46:49.4898684Z MAX_JOBS=94 2025-09-07T07:46:49.4898831Z GITHUB_ACTOR_ID=97764156 2025-09-07T07:46:49.4899149Z GITHUB_WORKFLOW_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:49.4899452Z GITHUB_REF_NAME=main 2025-09-07T07:46:49.4899704Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-09-07T07:46:49.4899974Z GITHUB_JOB=test 2025-09-07T07:46:49.4900125Z NO_TEST_TIMEOUT=False 2025-09-07T07:46:49.4900296Z TD_DISTRIBUTED=False 2025-09-07T07:46:49.4900478Z GITHUB_REPOSITORY=pytorch/pytorch 2025-09-07T07:46:49.4900682Z GITHUB_RETENTION_DAYS=90 2025-09-07T07:46:49.4900851Z OPENSSL_DIR=/opt/openssl 2025-09-07T07:46:49.4901030Z GITHUB_ACTION_REPOSITORY= 2025-09-07T07:46:49.4901531Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.9/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-09-07T07:46:49.4901990Z GITHUB_BASE_REF= 2025-09-07T07:46:49.4902143Z INSTALLED_ACL= 2025-09-07T07:46:49.4902444Z ARTIFACTS_FILE_SUFFIX=test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800 2025-09-07T07:46:49.4902770Z CI=true 2025-09-07T07:46:49.4902930Z GITHUB_REPOSITORY_OWNER=pytorch 2025-09-07T07:46:49.4903160Z RUST_LOG=sccache::server=error 2025-09-07T07:46:49.4903347Z JOB_ID=49775585800 2025-09-07T07:46:49.4903503Z GITHUB_HEAD_REF= 2025-09-07T07:46:49.4903665Z GITHUB_ACTION_REF= 2025-09-07T07:46:49.4903856Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-09-07T07:46:49.4904081Z TEST_SHOWLOCALS=False 2025-09-07T07:46:49.4904275Z GITHUB_WORKFLOW=inductor-perf-nightly-x86 2025-09-07T07:46:49.4904498Z DEBIAN_FRONTEND=noninteractive 2025-09-07T07:46:49.4904877Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_0a6d5ec5-039b-4965-a10d-8df077d2d803 2025-09-07T07:46:49.4905250Z NO_TD=False 2025-09-07T07:46:49.4905416Z SKIP_SCCACHE_INITIALIZATION=1 2025-09-07T07:46:49.4905623Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-09-07T07:46:49.4905819Z _=/usr/bin/env 2025-09-07T07:46:49.4906039Z ++ python -c 'import site; print(site.getsitepackages()[0])' 2025-09-07T07:46:49.5132549Z + TORCH_INSTALL_DIR=/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch 2025-09-07T07:46:49.5132913Z + TORCH_BIN_DIR=/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/bin 2025-09-07T07:46:49.5133258Z + TORCH_LIB_DIR=/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/lib 2025-09-07T07:46:49.5133594Z + TORCH_TEST_DIR=/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/test 2025-09-07T07:46:49.5133863Z + BUILD_DIR=build 2025-09-07T07:46:49.5134049Z + BUILD_RENAMED_DIR=build_renamed 2025-09-07T07:46:49.5134248Z + BUILD_BIN_DIR=build/bin 2025-09-07T07:46:49.5134410Z + SHARD_NUMBER=2 2025-09-07T07:46:49.5134567Z + NUM_TEST_SHARDS=4 2025-09-07T07:46:49.5134739Z + export TORCH_SERIALIZATION_DEBUG=1 2025-09-07T07:46:49.5134941Z + TORCH_SERIALIZATION_DEBUG=1 2025-09-07T07:46:49.5135111Z + export VALGRIND=ON 2025-09-07T07:46:49.5135274Z + VALGRIND=ON 2025-09-07T07:46:49.5135455Z + [[ linux-jammy-py3.9-gcc11-build == *clang9* ]] 2025-09-07T07:46:49.5135689Z + [[ linux-jammy-py3.9-gcc11-build == *xpu* ]] 2025-09-07T07:46:49.5135881Z + detect_cuda_arch 2025-09-07T07:46:49.5136065Z + [[ linux-jammy-py3.9-gcc11-build == *cuda* ]] 2025-09-07T07:46:49.5136296Z + [[ linux-jammy-py3.9-gcc11-build == *s390x* ]] 2025-09-07T07:46:49.5136495Z + [[ 0 == \1 ]] 2025-09-07T07:46:49.5136635Z + [[ True == \1 ]] 2025-09-07T07:46:49.5136814Z + [[ linux-jammy-py3.9-gcc11-build != *bazel* ]] 2025-09-07T07:46:49.5137032Z ++ realpath build/custom_test_artifacts 2025-09-07T07:46:49.5137334Z + CUSTOM_TEST_ARTIFACT_BUILD_DIR=/var/lib/jenkins/workspace/build/custom_test_artifacts 2025-09-07T07:46:49.5137622Z + [[ -n '' ]] 2025-09-07T07:46:49.5137770Z + echo 'Environment variables' 2025-09-07T07:46:49.5137949Z Environment variables 2025-09-07T07:46:49.5138101Z + env 2025-09-07T07:46:49.5155580Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T07:46:49.5155832Z CONTINUE_THROUGH_ERROR=True 2025-09-07T07:46:49.5156078Z BUILD_ENVIRONMENT=linux-jammy-py3.9-gcc11-build 2025-09-07T07:46:49.5156415Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-09-07T07:46:49.5156664Z HOSTNAME=5b1006e7d104 2025-09-07T07:46:49.5156999Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_0a6d5ec5-039b-4965-a10d-8df077d2d803 2025-09-07T07:46:49.5157354Z GITHUB_ACTION=__run_2 2025-09-07T07:46:49.5157519Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-09-07T07:46:49.5157707Z GITHUB_RUN_NUMBER=525 2025-09-07T07:46:49.5157893Z TEST_CONFIG=inductor_torchbench_perf_cpu_x86 2025-09-07T07:46:49.5158107Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-09-07T07:46:49.5158309Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-09-07T07:46:49.5158508Z SCCACHE_IDLE_TIMEOUT=0 2025-09-07T07:46:49.5158756Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-09-07T07:46:49.5158959Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-09-07T07:46:49.5159150Z GITHUB_REF_TYPE=branch 2025-09-07T07:46:49.5159343Z BASE_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:49.5159548Z XLA_CUDA= 2025-09-07T07:46:49.5159704Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-09-07T07:46:49.5159975Z HUGGING_FACE_HUB_TOKEN=*** 2025-09-07T07:46:49.5160169Z *** 2025-09-07T07:46:49.5160316Z GITHUB_REPOSITORY_ID=65600975 2025-09-07T07:46:49.5160495Z GITHUB_ACTIONS=true 2025-09-07T07:46:49.5160689Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-09-07T07:46:49.5160927Z SHA1=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:49.5161165Z GITHUB_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:49.5161551Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/inductor-perf-test-nightly-x86.yml@refs/heads/main 2025-09-07T07:46:49.5161897Z UCC_HOME=/usr 2025-09-07T07:46:49.5162049Z TORCH_SERIALIZATION_DEBUG=1 2025-09-07T07:46:49.5162226Z VERBOSE_TEST_LOGS=False 2025-09-07T07:46:49.5162402Z GITHUB_REF=refs/heads/main 2025-09-07T07:46:49.5162578Z SHARD_NUMBER=2 2025-09-07T07:46:49.5162733Z GITHUB_REF_PROTECTED=true 2025-09-07T07:46:49.5162904Z HOME=/var/lib/jenkins 2025-09-07T07:46:49.5163091Z GITHUB_API_URL=https://api.github.com 2025-09-07T07:46:49.5163304Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-09-07T07:46:49.5163486Z UCX_COMMIT= 2025-09-07T07:46:49.5163630Z USE_SYSTEM_NCCL=1 2025-09-07T07:46:49.5163786Z NUM_TEST_SHARDS=4 2025-09-07T07:46:49.5163938Z UCX_HOME=/usr 2025-09-07T07:46:49.5164264Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_0a6d5ec5-039b-4965-a10d-8df077d2d803 2025-09-07T07:46:49.5164796Z JOB_NAME=inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal) 2025-09-07T07:46:49.5165314Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_0a6d5ec5-039b-4965-a10d-8df077d2d803 2025-09-07T07:46:49.5165772Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-09-07T07:46:49.5166070Z GITHUB_EVENT_NAME=schedule 2025-09-07T07:46:49.5166496Z DASHBOARD_TAG=training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true 2025-09-07T07:46:49.5166937Z GITHUB_RUN_ID=17525285611 2025-09-07T07:46:49.5167110Z INSTALLED_OPENBLAS= 2025-09-07T07:46:49.5167475Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_0a6d5ec5-039b-4965-a10d-8df077d2d803 2025-09-07T07:46:49.5167860Z GITHUB_ACTOR=pytorchmergebot 2025-09-07T07:46:49.5168041Z PR_NUMBER= 2025-09-07T07:46:49.5168189Z DESIRED_CUDA= 2025-09-07T07:46:49.5168393Z GITHUB_RUN_ATTEMPT=1 2025-09-07T07:46:49.5168554Z VALGRIND=ON 2025-09-07T07:46:49.5168712Z ANACONDA_PYTHON_VERSION=3.9 2025-09-07T07:46:49.5168935Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-09-07T07:46:49.5169168Z TERM=vt100 2025-09-07T07:46:49.5169307Z INSTALLED_VISION=yes 2025-09-07T07:46:49.5169470Z BRANCH=main 2025-09-07T07:46:49.5169622Z SCCACHE_REGION=us-east-1 2025-09-07T07:46:49.5169862Z OPENSSL_ROOT_DIR=/opt/openssl 2025-09-07T07:46:49.5170040Z CUDA_PATH=/usr/local/cuda 2025-09-07T07:46:49.5170350Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-09-07T07:46:49.5170690Z GITHUB_SERVER_URL=https://github.com 2025-09-07T07:46:49.5170934Z UCC_COMMIT= 2025-09-07T07:46:49.5171070Z REENABLED_ISSUES= 2025-09-07T07:46:49.5171220Z DOCS=yes 2025-09-07T07:46:49.5171361Z SHLVL=1 2025-09-07T07:46:49.5171518Z MAX_JOBS=94 2025-09-07T07:46:49.5171655Z GITHUB_ACTOR_ID=97764156 2025-09-07T07:46:49.5171883Z GITHUB_WORKFLOW_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:49.5172124Z GITHUB_REF_NAME=main 2025-09-07T07:46:49.5172366Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-09-07T07:46:49.5172617Z GITHUB_JOB=test 2025-09-07T07:46:49.5172777Z NO_TEST_TIMEOUT=False 2025-09-07T07:46:49.5172942Z TD_DISTRIBUTED=False 2025-09-07T07:46:49.5173119Z GITHUB_REPOSITORY=pytorch/pytorch 2025-09-07T07:46:49.5173313Z GITHUB_RETENTION_DAYS=90 2025-09-07T07:46:49.5173489Z OPENSSL_DIR=/opt/openssl 2025-09-07T07:46:49.5173665Z GITHUB_ACTION_REPOSITORY= 2025-09-07T07:46:49.5174118Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.9/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-09-07T07:46:49.5174566Z GITHUB_BASE_REF= 2025-09-07T07:46:49.5174727Z INSTALLED_ACL= 2025-09-07T07:46:49.5175017Z ARTIFACTS_FILE_SUFFIX=test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800 2025-09-07T07:46:49.5175336Z CI=true 2025-09-07T07:46:49.5175481Z GITHUB_REPOSITORY_OWNER=pytorch 2025-09-07T07:46:49.5175696Z RUST_LOG=sccache::server=error 2025-09-07T07:46:49.5175872Z JOB_ID=49775585800 2025-09-07T07:46:49.5176025Z GITHUB_HEAD_REF= 2025-09-07T07:46:49.5176176Z GITHUB_ACTION_REF= 2025-09-07T07:46:49.5176374Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-09-07T07:46:49.5176599Z TEST_SHOWLOCALS=False 2025-09-07T07:46:49.5176794Z GITHUB_WORKFLOW=inductor-perf-nightly-x86 2025-09-07T07:46:49.5177011Z DEBIAN_FRONTEND=noninteractive 2025-09-07T07:46:49.5177387Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_0a6d5ec5-039b-4965-a10d-8df077d2d803 2025-09-07T07:46:49.5177767Z NO_TD=False 2025-09-07T07:46:49.5177931Z SKIP_SCCACHE_INITIALIZATION=1 2025-09-07T07:46:49.5178127Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-09-07T07:46:49.5178326Z _=/usr/bin/env 2025-09-07T07:46:49.5178487Z + echo 'Testing pytorch' 2025-09-07T07:46:49.5178664Z Testing pytorch 2025-09-07T07:46:49.5178824Z + export LANG=C.UTF-8 2025-09-07T07:46:49.5178989Z + LANG=C.UTF-8 2025-09-07T07:46:49.5179143Z + PR_NUMBER= 2025-09-07T07:46:49.5179345Z + [[ inductor_torchbench_perf_cpu_x86 == \d\e\f\a\u\l\t ]] 2025-09-07T07:46:49.5179630Z + [[ inductor_torchbench_perf_cpu_x86 == \d\i\s\t\r\i\b\u\t\e\d ]] 2025-09-07T07:46:49.5179912Z + [[ inductor_torchbench_perf_cpu_x86 == \s\l\o\w ]] 2025-09-07T07:46:49.5180191Z + [[ linux-jammy-py3.9-gcc11-build == *slow-gradcheck* ]] 2025-09-07T07:46:49.5180457Z + [[ linux-jammy-py3.9-gcc11-build == *cuda* ]] 2025-09-07T07:46:49.5180694Z + [[ linux-jammy-py3.9-gcc11-build == *rocm* ]] 2025-09-07T07:46:49.5180922Z + [[ linux-jammy-py3.9-gcc11-build == *xpu* ]] 2025-09-07T07:46:49.5181174Z + [[ inductor_torchbench_perf_cpu_x86 == *crossref* ]] 2025-09-07T07:46:49.5181438Z + [[ linux-jammy-py3.9-gcc11-build == *rocm* ]] 2025-09-07T07:46:49.5181669Z + [[ linux-jammy-py3.9-gcc11-build == *xpu* ]] 2025-09-07T07:46:49.5181907Z + [[ linux-jammy-py3.9-gcc11-build != *-bazel-* ]] 2025-09-07T07:46:49.5182140Z + pip_install ninja==1.10.2 2025-09-07T07:46:49.5182386Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-09-07T07:46:49.5182680Z + python3 -m pip install --progress-bar off ninja==1.10.2 2025-09-07T07:46:49.8688015Z Collecting ninja==1.10.2 2025-09-07T07:46:49.8776684Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl.metadata (5.0 kB) 2025-09-07T07:46:49.8868614Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl (108 kB) 2025-09-07T07:46:50.5689124Z Installing collected packages: ninja 2025-09-07T07:46:50.5689377Z Attempting uninstall: ninja 2025-09-07T07:46:50.5689701Z Found existing installation: ninja 1.11.1.3 2025-09-07T07:46:50.5704346Z Uninstalling ninja-1.11.1.3: 2025-09-07T07:46:50.5744913Z Successfully uninstalled ninja-1.11.1.3 2025-09-07T07:46:50.6158084Z Successfully installed ninja-1.10.2 2025-09-07T07:46:50.6952358Z + export PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.9/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-09-07T07:46:50.6953207Z + PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.9/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-09-07T07:46:50.6953771Z + [[ linux-jammy-py3.9-gcc11-build == *aarch64* ]] 2025-09-07T07:46:50.6954030Z + [[ linux-jammy-py3.9-gcc11-build == *asan* ]] 2025-09-07T07:46:50.6954269Z + [[ linux-jammy-py3.9-gcc11-build == *-debug* ]] 2025-09-07T07:46:50.6954508Z + [[ linux-jammy-py3.9-gcc11-build != *-bazel-* ]] 2025-09-07T07:46:50.6954822Z + echo 'We are not in debug mode: linux-jammy-py3.9-gcc11-build. Expect the assertion to pass' 2025-09-07T07:46:50.6955222Z We are not in debug mode: linux-jammy-py3.9-gcc11-build. Expect the assertion to pass 2025-09-07T07:46:50.6955487Z + cd test 2025-09-07T07:46:50.6955711Z + python -c 'import torch; torch._C._crash_if_debug_asserts_fail(424242)' 2025-09-07T07:46:50.9778960Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:46:50.9780499Z import pynvml # type: ignore[import] 2025-09-07T07:46:51.7321789Z + [[ inductor_torchbench_perf_cpu_x86 == \n\o\g\p\u\_\N\O\_\A\V\X\2 ]] 2025-09-07T07:46:51.7322212Z + [[ inductor_torchbench_perf_cpu_x86 == \n\o\g\p\u\_\A\V\X\5\1\2 ]] 2025-09-07T07:46:51.7322586Z + [[ inductor_torchbench_perf_cpu_x86 == \l\e\g\a\c\y\_\n\v\i\d\i\a\_\d\r\i\v\e\r ]] 2025-09-07T07:46:51.7323304Z + DYNAMO_BENCHMARK_FLAGS=() 2025-09-07T07:46:51.7325077Z + [[ inductor_torchbench_perf_cpu_x86 == *pr_time_benchmarks* ]] 2025-09-07T07:46:51.7325405Z + [[ inductor_torchbench_perf_cpu_x86 == *dynamo_eager* ]] 2025-09-07T07:46:51.7325764Z + [[ inductor_torchbench_perf_cpu_x86 == *aot_eager* ]] 2025-09-07T07:46:51.7326061Z + [[ inductor_torchbench_perf_cpu_x86 == *aot_inductor* ]] 2025-09-07T07:46:51.7326389Z + [[ inductor_torchbench_perf_cpu_x86 == *max_autotune_inductor* ]] 2025-09-07T07:46:51.7326712Z + [[ inductor_torchbench_perf_cpu_x86 == *inductor* ]] 2025-09-07T07:46:51.7326999Z + [[ inductor_torchbench_perf_cpu_x86 != *perf* ]] 2025-09-07T07:46:51.7327286Z + [[ inductor_torchbench_perf_cpu_x86 == *dynamic* ]] 2025-09-07T07:46:51.7327565Z + [[ inductor_torchbench_perf_cpu_x86 == *cpu* ]] 2025-09-07T07:46:51.7327840Z + DYNAMO_BENCHMARK_FLAGS+=(--device cpu) 2025-09-07T07:46:51.7752365Z + [[ linux-jammy-py3.9-gcc11-build == *libtorch* ]] 2025-09-07T07:46:51.7752864Z + [[ linux-jammy-py3.9-gcc11-build == *-bazel-* ]] 2025-09-07T07:46:51.7754285Z + cd test 2025-09-07T07:46:51.7754605Z + python -c 'import torch; print(torch.__config__.show())' 2025-09-07T07:46:52.0487925Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:46:52.0489320Z import pynvml # type: ignore[import] 2025-09-07T07:46:52.6368616Z PyTorch built with: 2025-09-07T07:46:52.6368908Z - GCC 11.4 2025-09-07T07:46:52.6369102Z - C++ Version: 201703 2025-09-07T07:46:52.6369859Z - Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-09-07T07:46:52.6370337Z - Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-09-07T07:46:52.6370644Z - OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-09-07T07:46:52.6371020Z - LAPACK is enabled (usually provided by MKL) 2025-09-07T07:46:52.6371269Z - NNPACK is enabled 2025-09-07T07:46:52.6371463Z - CPU capability usage: AVX512 2025-09-07T07:46:52.6373944Z - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, COMMIT_SHA=93fb23d6fae7c4e82c4239a1033e522088742634, CXX_COMPILER=/opt/cache/bin/c++, CXX_FLAGS= -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOCUPTI -DLIBKINETO_NOROCTRACER -DLIBKINETO_NOXPUPTI=ON -DUSE_FBGEMM -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -DC10_NODEPRECATED -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-unknown-pragmas -Wno-unused-parameter -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=old-style-cast -faligned-new -Werror -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, TORCH_VERSION=2.9.0, USE_CUDA=OFF, USE_CUDNN=OFF, USE_CUSPARSELT=OFF, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=OFF, USE_NCCL=OFF, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, USE_ROCM_KERNEL_ASSERT=OFF, USE_XCCL=OFF, USE_XPU=OFF, 2025-09-07T07:46:52.6376432Z 2025-09-07T07:46:52.8034921Z + cd test 2025-09-07T07:46:52.8035278Z + python -c 'import torch; print(torch.__config__.parallel_info())' 2025-09-07T07:46:53.0792555Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:46:53.0793949Z import pynvml # type: ignore[import] 2025-09-07T07:46:53.6558551Z ATen/Parallel: 2025-09-07T07:46:53.6558802Z at::get_num_threads() : 48 2025-09-07T07:46:53.6559050Z at::get_num_interop_threads() : 48 2025-09-07T07:46:53.6559260Z OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-09-07T07:46:53.6559458Z omp_get_max_threads() : 48 2025-09-07T07:46:53.6559784Z Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-09-07T07:46:53.6560121Z mkl_get_max_threads() : 48 2025-09-07T07:46:53.6560371Z Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-09-07T07:46:53.6560637Z std::thread::hardware_concurrency() : 96 2025-09-07T07:46:53.6560839Z Environment variables: 2025-09-07T07:46:53.6561008Z OMP_NUM_THREADS : [not set] 2025-09-07T07:46:53.6561183Z MKL_NUM_THREADS : [not set] 2025-09-07T07:46:53.6561374Z ATen parallel backend: OpenMP 2025-09-07T07:46:53.6561490Z 2025-09-07T07:46:53.8211491Z + [[ inductor_torchbench_perf_cpu_x86 == *numpy_2* ]] 2025-09-07T07:46:53.8211867Z + [[ linux-jammy-py3.9-gcc11-build == *aarch64* ]] 2025-09-07T07:46:53.8212180Z + [[ inductor_torchbench_perf_cpu_x86 == *backward* ]] 2025-09-07T07:46:53.8212491Z + [[ inductor_torchbench_perf_cpu_x86 == *xla* ]] 2025-09-07T07:46:53.8212765Z + [[ inductor_torchbench_perf_cpu_x86 == *vllm* ]] 2025-09-07T07:46:53.8213054Z + [[ inductor_torchbench_perf_cpu_x86 == *executorch* ]] 2025-09-07T07:46:53.8213361Z + [[ inductor_torchbench_perf_cpu_x86 == \j\i\t\_\l\e\g\a\c\y ]] 2025-09-07T07:46:53.8213681Z + [[ linux-jammy-py3.9-gcc11-build == *libtorch* ]] 2025-09-07T07:46:53.8213971Z + [[ inductor_torchbench_perf_cpu_x86 == distributed ]] 2025-09-07T07:46:53.8214279Z + [[ inductor_torchbench_perf_cpu_x86 == *operator_benchmark* ]] 2025-09-07T07:46:53.8214623Z + [[ inductor_torchbench_perf_cpu_x86 == *inductor_distributed* ]] 2025-09-07T07:46:53.8215244Z + [[ inductor_torchbench_perf_cpu_x86 == *inductor-halide* ]] 2025-09-07T07:46:53.8215592Z + [[ inductor_torchbench_perf_cpu_x86 == *inductor-triton-cpu* ]] 2025-09-07T07:46:53.8215956Z + [[ inductor_torchbench_perf_cpu_x86 == *inductor-micro-benchmark* ]] 2025-09-07T07:46:53.8216399Z + [[ inductor_torchbench_perf_cpu_x86 == *huggingface* ]] 2025-09-07T07:46:53.8216676Z + [[ inductor_torchbench_perf_cpu_x86 == *timm* ]] 2025-09-07T07:46:53.8216950Z + [[ inductor_torchbench_perf_cpu_x86 == cachebench ]] 2025-09-07T07:46:53.8217248Z + [[ inductor_torchbench_perf_cpu_x86 == verify_cachebench ]] 2025-09-07T07:46:53.8217554Z + [[ inductor_torchbench_perf_cpu_x86 == *torchbench* ]] 2025-09-07T07:46:53.8217803Z + install_torchaudio 2025-09-07T07:46:53.8217994Z + local commit 2025-09-07T07:46:53.8218190Z ++ get_pinned_commit audio 2025-09-07T07:46:53.8218405Z ++ cat .github/ci_commit_pins/audio.txt 2025-09-07T07:46:53.8669539Z + commit=2e300559e4e123928a22187b8f59a5b56f57ddc8 2025-09-07T07:46:53.8670505Z + pip_build_and_install git+https://github.com/pytorch/audio.git@2e300559e4e123928a22187b8f59a5b56f57ddc8 dist/audio 2025-09-07T07:46:53.8671075Z + local build_target=git+https://github.com/pytorch/audio.git@2e300559e4e123928a22187b8f59a5b56f57ddc8 2025-09-07T07:46:53.8671433Z + local wheel_dir=dist/audio 2025-09-07T07:46:53.8671663Z + local found_whl=0 2025-09-07T07:46:53.8671855Z + for file in "${wheel_dir}"/*.whl 2025-09-07T07:46:53.8672151Z + [[ -f dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl ]] 2025-09-07T07:46:53.8672420Z + found_whl=1 2025-09-07T07:46:53.8672568Z + break 2025-09-07T07:46:53.8672709Z + '[' 1 == 0 ']' 2025-09-07T07:46:53.8672867Z + for file in "${wheel_dir}"/*.whl 2025-09-07T07:46:53.8673153Z + pip_install_whl dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl 2025-09-07T07:46:53.8673517Z + args=('dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl') 2025-09-07T07:46:53.8673771Z + local args 2025-09-07T07:46:53.8674005Z + [[ dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl == *\ * ]] 2025-09-07T07:46:53.8674288Z + for path in "${args[@]}" 2025-09-07T07:46:53.8674562Z + echo 'Installing dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl' 2025-09-07T07:46:53.8674930Z Installing dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl 2025-09-07T07:46:53.8675354Z + python3 -mpip install --no-index --no-deps dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl 2025-09-07T07:46:54.1204804Z Processing ./dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl 2025-09-07T07:46:54.1241552Z Installing collected packages: torchaudio 2025-09-07T07:46:54.8926657Z Successfully installed torchaudio-2.8.0a0+2e30055 2025-09-07T07:46:54.9245108Z + install_torchvision 2025-09-07T07:46:54.9245310Z + local orig_preload 2025-09-07T07:46:54.9245476Z + local commit 2025-09-07T07:46:54.9245632Z ++ get_pinned_commit vision 2025-09-07T07:46:54.9245829Z ++ cat .github/ci_commit_pins/vision.txt 2025-09-07T07:46:54.9250044Z + commit=966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-09-07T07:46:54.9250349Z + orig_preload= 2025-09-07T07:46:54.9250532Z + '[' -n '' ']' 2025-09-07T07:46:54.9250724Z + [[ linux-jammy-py3.9-gcc11-build == *cuda* ]] 2025-09-07T07:46:54.9251291Z + pip_build_and_install git+https://github.com/pytorch/vision.git@966da7e46f65d6d49df3e31214470a4fe5cc8e66 dist/vision 2025-09-07T07:46:54.9251895Z + local build_target=git+https://github.com/pytorch/vision.git@966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-09-07T07:46:54.9252234Z + local wheel_dir=dist/vision 2025-09-07T07:46:54.9252420Z + local found_whl=0 2025-09-07T07:46:54.9252587Z + for file in "${wheel_dir}"/*.whl 2025-09-07T07:46:54.9252866Z + [[ -f dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl ]] 2025-09-07T07:46:54.9253138Z + found_whl=1 2025-09-07T07:46:54.9253286Z + break 2025-09-07T07:46:54.9253419Z + '[' 1 == 0 ']' 2025-09-07T07:46:54.9253570Z + for file in "${wheel_dir}"/*.whl 2025-09-07T07:46:54.9254077Z + pip_install_whl dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl 2025-09-07T07:46:54.9254465Z + args=('dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl') 2025-09-07T07:46:54.9254730Z + local args 2025-09-07T07:46:54.9254971Z + [[ dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl == *\ * ]] 2025-09-07T07:46:54.9255353Z + for path in "${args[@]}" 2025-09-07T07:46:54.9255641Z + echo 'Installing dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl' 2025-09-07T07:46:54.9256028Z Installing dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl 2025-09-07T07:46:54.9256468Z + python3 -mpip install --no-index --no-deps dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl 2025-09-07T07:46:55.1740301Z Processing ./dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl 2025-09-07T07:46:55.1797480Z Installing collected packages: torchvision 2025-09-07T07:46:55.7502856Z Successfully installed torchvision-0.22.0a0+966da7e 2025-09-07T07:46:55.7809428Z + '[' -n '' ']' 2025-09-07T07:46:55.7809636Z + id=1 2025-09-07T07:46:55.7809840Z + pip_install opencv-python==4.8.0.74 2025-09-07T07:46:55.7810164Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-09-07T07:46:55.7810508Z + python3 -m pip install --progress-bar off opencv-python==4.8.0.74 2025-09-07T07:46:56.0962416Z Collecting opencv-python==4.8.0.74 2025-09-07T07:46:56.1217064Z Downloading opencv_python-4.8.0.74-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (19 kB) 2025-09-07T07:46:56.1294025Z Requirement already satisfied: numpy>=1.17.0 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from opencv-python==4.8.0.74) (1.22.4) 2025-09-07T07:46:56.1397527Z Downloading opencv_python-4.8.0.74-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (61.7 MB) 2025-09-07T07:46:57.3147912Z Installing collected packages: opencv-python 2025-09-07T07:46:57.3148304Z Attempting uninstall: opencv-python 2025-09-07T07:46:57.3149632Z Found existing installation: opencv-python 4.11.0.86 2025-09-07T07:46:57.3201526Z Uninstalling opencv-python-4.11.0.86: 2025-09-07T07:46:57.3956644Z Successfully uninstalled opencv-python-4.11.0.86 2025-09-07T07:46:58.1404892Z Successfully installed opencv-python-4.8.0.74 2025-09-07T07:46:58.2277833Z + [[ inductor_torchbench_perf_cpu_x86 == *inductor_torchbench_smoketest_perf* ]] 2025-09-07T07:46:58.2278348Z + [[ inductor_torchbench_perf_cpu_x86 == *inductor_torchbench_cpu_smoketest_perf* ]] 2025-09-07T07:46:58.2278751Z + [[ inductor_torchbench_perf_cpu_x86 == *torchbench_gcp_smoketest* ]] 2025-09-07T07:46:58.2279076Z + [[ inductor_torchbench_perf_cpu_x86 != *cpu* ]] 2025-09-07T07:46:58.2279309Z + PYTHONPATH=/torchbench 2025-09-07T07:46:58.2279507Z + test_dynamo_benchmark torchbench 1 2025-09-07T07:46:58.2279722Z ++ pwd 2025-09-07T07:46:58.2279940Z + TEST_REPORTS_DIR=/var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:46:58.2280186Z + local suite=torchbench 2025-09-07T07:46:58.2280350Z + shift 2025-09-07T07:46:58.2280508Z + local shard_id=1 2025-09-07T07:46:58.2280665Z + shift 2025-09-07T07:46:58.2280847Z + [[ inductor_torchbench_perf_cpu_x86 == *perf_compare* ]] 2025-09-07T07:46:58.2281102Z + [[ inductor_torchbench_perf_cpu_x86 == *perf* ]] 2025-09-07T07:46:58.2281319Z + [[ inductor_torchbench_perf_cpu_x86 == *b200* ]] 2025-09-07T07:46:58.2281562Z + test_single_dynamo_benchmark dashboard torchbench 1 2025-09-07T07:46:58.2281772Z ++ pwd 2025-09-07T07:46:58.2281966Z + TEST_REPORTS_DIR=/var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:46:58.2282236Z + mkdir -p /var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:46:58.2294653Z + local name=dashboard 2025-09-07T07:46:58.2294813Z + shift 2025-09-07T07:46:58.2294953Z + local suite=torchbench 2025-09-07T07:46:58.2295116Z + shift 2025-09-07T07:46:58.2295239Z + local shard_id=1 2025-09-07T07:46:58.2295385Z + shift 2025-09-07T07:46:58.2295535Z + partition_flags=() 2025-09-07T07:46:58.2295701Z + local partition_flags 2025-09-07T07:46:58.2296093Z + [[ -n 4 ]] 2025-09-07T07:46:58.2296243Z + [[ -n 1 ]] 2025-09-07T07:46:58.2296494Z + partition_flags=(--total-partitions "$NUM_TEST_SHARDS" --partition-id "$shard_id") 2025-09-07T07:46:58.2296815Z + [[ inductor_torchbench_perf_cpu_x86 == *perf_compare* ]] 2025-09-07T07:46:58.2297052Z + [[ inductor_torchbench_perf_cpu_x86 == *perf* ]] 2025-09-07T07:46:58.2297456Z + test_perf_for_dashboard torchbench --device cpu --total-partitions 4 --partition-id 1 2025-09-07T07:46:58.2298929Z ++ pwd 2025-09-07T07:46:58.2300441Z + TEST_REPORTS_DIR=/var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:46:58.2300721Z + mkdir -p /var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:46:58.2312266Z + local suite=torchbench 2025-09-07T07:46:58.2312430Z + shift 2025-09-07T07:46:58.2312573Z + local backend=inductor 2025-09-07T07:46:58.2312730Z + modes=() 2025-09-07T07:46:58.2312877Z + local modes 2025-09-07T07:46:58.2313341Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true == *training-true* ]] 2025-09-07T07:46:58.2314102Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true == *inference-true* ]] 2025-09-07T07:46:58.2314574Z + modes+=(inference) 2025-09-07T07:46:58.2314746Z + targets=('accuracy' 'performance') 2025-09-07T07:46:58.2314942Z + local targets 2025-09-07T07:46:58.2315093Z + local device=cuda 2025-09-07T07:46:58.2315271Z + [[ inductor_torchbench_perf_cpu_x86 == *cpu* ]] 2025-09-07T07:46:58.2315511Z + [[ inductor_torchbench_perf_cpu_x86 == *cpu_x86_zen* ]] 2025-09-07T07:46:58.2315762Z + [[ inductor_torchbench_perf_cpu_x86 == *cpu_x86* ]] 2025-09-07T07:46:58.2315975Z + device=cpu_x86 2025-09-07T07:46:58.2316140Z + test_inductor_set_cpu_affinity 2025-09-07T07:46:58.2316642Z ++ find /usr/lib -name libjemalloc.so.2 2025-09-07T07:46:58.2495552Z + JEMALLOC_LIB=/usr/lib/x86_64-linux-gnu/libjemalloc.so.2 2025-09-07T07:46:58.2495837Z + export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libjemalloc.so.2: 2025-09-07T07:46:58.2496128Z + LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libjemalloc.so.2: 2025-09-07T07:46:58.2496533Z + export MALLOC_CONF=oversize_threshold:1,background_thread:true,metadata_thp:auto,dirty_decay_ms:-1,muzzy_decay_ms:-1 2025-09-07T07:46:58.2497027Z + MALLOC_CONF=oversize_threshold:1,background_thread:true,metadata_thp:auto,dirty_decay_ms:-1,muzzy_decay_ms:-1 2025-09-07T07:46:58.2497403Z + [[ inductor_torchbench_perf_cpu_x86 != *aarch64* ]] 2025-09-07T07:46:58.2497996Z +++ which python 2025-09-07T07:46:58.2513518Z ++ dirname /opt/conda/envs/py_3.9/bin/python 2025-09-07T07:46:58.2542301Z + IOMP_LIB=/opt/conda/envs/py_3.9/bin/../lib/libiomp5.so 2025-09-07T07:46:58.2542687Z + export LD_PRELOAD=/opt/conda/envs/py_3.9/bin/../lib/libiomp5.so:/usr/lib/x86_64-linux-gnu/libjemalloc.so.2: 2025-09-07T07:46:58.2543147Z + LD_PRELOAD=/opt/conda/envs/py_3.9/bin/../lib/libiomp5.so:/usr/lib/x86_64-linux-gnu/libjemalloc.so.2: 2025-09-07T07:46:58.2543494Z + export KMP_AFFINITY=granularity=fine,compact,1,0 2025-09-07T07:46:58.2543742Z + KMP_AFFINITY=granularity=fine,compact,1,0 2025-09-07T07:46:58.2543963Z + export KMP_BLOCKTIME=1 2025-09-07T07:46:58.2544134Z + KMP_BLOCKTIME=1 2025-09-07T07:46:58.2546692Z ++ nproc 2025-09-07T07:46:58.2566435Z + cpus=96 2025-09-07T07:46:58.2573253Z ++ lscpu 2025-09-07T07:46:58.2574466Z ++ grep 'Thread(s) per core:' 2025-09-07T07:46:58.2575140Z ++ awk '{print $4}' 2025-09-07T07:46:58.2830833Z + thread_per_core=2 2025-09-07T07:46:58.2831056Z + cores=48 2025-09-07T07:46:58.2831258Z + [[ inductor_torchbench_perf_cpu_x86 == *aarch64* ]] 2025-09-07T07:46:58.2831505Z + export OMP_NUM_THREADS=48 2025-09-07T07:46:58.2831692Z + OMP_NUM_THREADS=48 2025-09-07T07:46:58.2835489Z ++ python -c 'import os; print(min(os.sched_getaffinity(0)))' 2025-09-07T07:46:58.3069684Z + start_cpu=0 2025-09-07T07:46:58.3070039Z ++ python -c 'import os; print(max(os.sched_getaffinity(0)))' 2025-09-07T07:46:58.3282864Z + end_cpu=93 2025-09-07T07:46:58.3283223Z + export 'TASKSET=taskset -c 0-93' 2025-09-07T07:46:58.3283743Z + TASKSET='taskset -c 0-93' 2025-09-07T07:46:58.3283965Z + for mode in "${modes[@]}" 2025-09-07T07:46:58.3284167Z + [[ inference == \i\n\f\e\r\e\n\c\e ]] 2025-09-07T07:46:58.3284389Z + [[ cpu_x86 == \c\p\u\_\x\8\6 ]] 2025-09-07T07:46:58.3284583Z + dtype=amp 2025-09-07T07:46:58.3284749Z + for target in "${targets[@]}" 2025-09-07T07:46:58.3285019Z + target_flag=('--accuracy') 2025-09-07T07:46:58.3285207Z + local target_flag 2025-09-07T07:46:58.3285391Z + [[ accuracy == \p\e\r\f\o\r\m\a\n\c\e ]] 2025-09-07T07:46:58.3285609Z + [[ accuracy == \a\c\c\u\r\a\c\y ]] 2025-09-07T07:46:58.3285832Z + target_flag+=(--no-translation-validation) 2025-09-07T07:46:58.3286375Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true == *freezing-true* ]] 2025-09-07T07:46:58.3286872Z + target_flag+=(--freezing) 2025-09-07T07:46:58.3287348Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true == *default-true* ]] 2025-09-07T07:46:58.3288549Z + taskset -c 0-93 python benchmarks/dynamo/torchbench.py --accuracy --no-translation-validation --freezing --inference --amp --backend inductor --disable-cudagraphs --device cpu --total-partitions 4 --partition-id 1 --output /var/lib/jenkins/workspace/test/test-reports/inductor_no_cudagraphs_torchbench_amp_inference_cpu_x86_accuracy.csv 2025-09-07T07:46:58.6490908Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:46:58.6492301Z import pynvml # type: ignore[import] 2025-09-07T07:47:00.7636227Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:47:00.7637653Z import pynvml # type: ignore[import] 2025-09-07T07:47:02.4759842Z 2025-09-07T07:47:07.5727840Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:47:07.5730257Z loading model: 0it [00:05, ?it/s] 2025-09-07T07:47:07.5731481Z cpu eval dlrm 2025-09-07T07:47:07.7635033Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:47:07.8130208Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:47:07.8611999Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:47:26.0990763Z pass 2025-09-07T07:47:26.0991240Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:47:27.7417656Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:47:27.7419056Z import pynvml # type: ignore[import] 2025-09-07T07:47:29.4537239Z 2025-09-07T07:47:30.4684006Z loading model: 0it [00:00, ?it/s]Downloading https://doctr-static.mindee.com/models?id=v0.7.0/db_resnet50-79bd7d70.pt&src=0 to /var/lib/jenkins/.cache/doctr/models/db_resnet50-79bd7d70.pt 2025-09-07T07:47:30.7771979Z 2025-09-07T07:47:30.7772075Z 2025-09-07T07:47:30.8772306Z 0% 0/102021912 [00:00 2025-09-07T08:00:01.0357613Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0357805Z 2025-09-07T08:00:01.0357905Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0358352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0358771Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0359089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0359408Z x = module(x) 2025-09-07T08:00:01.0359705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0360030Z out = self.relu(out) 2025-09-07T08:00:01.0360131Z 2025-09-07T08:00:01.0360236Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0360669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0361086Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0361494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:00:01.0361999Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0362489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:00:01.0362989Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0363180Z 2025-09-07T08:00:01.0363281Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0363722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0364135Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0364462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0364772Z x = module(x) 2025-09-07T08:00:01.0365069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0365396Z out = self.relu(out) 2025-09-07T08:00:01.0365496Z 2025-09-07T08:00:01.0365601Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0366034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0366449Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0366771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0367084Z x = module(x) 2025-09-07T08:00:01.0367378Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0367690Z out = self.relu(out) 2025-09-07T08:00:01.0367796Z 2025-09-07T08:00:01.0367897Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0368445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0368872Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0369196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0369510Z x = module(x) 2025-09-07T08:00:01.0369804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0370129Z out = self.relu(out) 2025-09-07T08:00:01.0370229Z 2025-09-07T08:00:01.0370315Z cudagraph partition due to non gpu ops 2025-09-07T08:00:01.0370544Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0370991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0371407Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0371729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0372042Z x = module(x) 2025-09-07T08:00:01.0372325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0372646Z out = self.relu(out) 2025-09-07T08:00:01.0372747Z 2025-09-07T08:00:01.0372851Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0373294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0373699Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0374018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0374334Z x = module(x) 2025-09-07T08:00:01.0374625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0375010Z out = self.relu(out) 2025-09-07T08:00:01.0375110Z 2025-09-07T08:00:01.0375210Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0375653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0376112Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0376435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0376755Z x = module(x) 2025-09-07T08:00:01.0377046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0377374Z out = self.relu(out) 2025-09-07T08:00:01.0377474Z 2025-09-07T08:00:01.0377582Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0378027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0378431Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0378749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0379066Z x = module(x) 2025-09-07T08:00:01.0379357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0379681Z out = self.relu(out) 2025-09-07T08:00:01.0379781Z 2025-09-07T08:00:01.0379859Z cudagraph partition due to non gpu ops 2025-09-07T08:00:01.0380091Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0380528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0380940Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0381251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0381572Z x = module(x) 2025-09-07T08:00:01.0381878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0382202Z out = self.relu(out) 2025-09-07T08:00:01.0382300Z 2025-09-07T08:00:01.0382408Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0382845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0383260Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0383578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0383891Z x = module(x) 2025-09-07T08:00:01.0384175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0384498Z out = self.relu(out) 2025-09-07T08:00:01.0384603Z 2025-09-07T08:00:01.0384704Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0385139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0385550Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0385866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0386182Z x = module(x) 2025-09-07T08:00:01.0386472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0386792Z out = self.relu(out) 2025-09-07T08:00:01.0386889Z 2025-09-07T08:00:01.0386993Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0387423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0387830Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0388174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0388495Z x = module(x) 2025-09-07T08:00:01.0388780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0389130Z out = self.relu(out) 2025-09-07T08:00:01.0389233Z 2025-09-07T08:00:01.0389309Z cudagraph partition due to non gpu ops 2025-09-07T08:00:01.0389543Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0389984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0390391Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0390713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0391033Z x = module(x) 2025-09-07T08:00:01.0391326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0391643Z out = self.relu(out) 2025-09-07T08:00:01.0391749Z 2025-09-07T08:00:01.0391850Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0392298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0392714Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0393037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0393346Z x = module(x) 2025-09-07T08:00:01.0393641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0393966Z out = self.relu(out) 2025-09-07T08:00:01.0394065Z 2025-09-07T08:00:01.0394173Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0394621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0395029Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0395354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0395674Z x = module(x) 2025-09-07T08:00:01.0395972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0396293Z out = self.relu(out) 2025-09-07T08:00:01.0396400Z 2025-09-07T08:00:01.0396501Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0396946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0397361Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0397680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0397990Z x = module(x) 2025-09-07T08:00:01.0398286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0398609Z out = self.relu(out) 2025-09-07T08:00:01.0398712Z 2025-09-07T08:00:01.0398818Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0399252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0399666Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0399990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0400308Z x = module(x) 2025-09-07T08:00:01.0400604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:00:01.0400968Z out = self.relu(out) 2025-09-07T08:00:01.0401079Z 2025-09-07T08:00:01.0401179Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0401630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0402076Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0402392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0402700Z x = module(x) 2025-09-07T08:00:01.0402994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:00:01.0403312Z out = self.relu(out) 2025-09-07T08:00:01.0403410Z 2025-09-07T08:00:01.0403493Z cudagraph partition due to non gpu ops 2025-09-07T08:00:01.0403719Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0404162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0404576Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0404987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:00:01.0405471Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0405951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:00:01.0406433Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0406629Z 2025-09-07T08:00:01.0406728Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0407169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0407582Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0407906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0408214Z x = module(x) 2025-09-07T08:00:01.0408504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0408825Z out = self.relu(out) 2025-09-07T08:00:01.0408921Z 2025-09-07T08:00:01.0409026Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0409448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0409848Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0410247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:00:01.0410708Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0411176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:00:01.0411635Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0411827Z 2025-09-07T08:00:01.0411926Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0412360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0412769Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0413073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0413384Z x = module(x) 2025-09-07T08:00:01.0413675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0414020Z out = self.relu(out) 2025-09-07T08:00:01.0414120Z 2025-09-07T08:00:01.0414224Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0414648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0415081Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0415397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0415708Z x = module(x) 2025-09-07T08:00:01.0415994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0416306Z out = self.relu(out) 2025-09-07T08:00:01.0416415Z 2025-09-07T08:00:01.0416511Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0416945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0417352Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0417658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0417965Z x = module(x) 2025-09-07T08:00:01.0418249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0418570Z out = self.relu(out) 2025-09-07T08:00:01.0418667Z 2025-09-07T08:00:01.0418750Z cudagraph partition due to non gpu ops 2025-09-07T08:00:01.0418965Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0419396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0419799Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0420108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0420411Z x = module(x) 2025-09-07T08:00:01.0420695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0421015Z out = self.relu(out) 2025-09-07T08:00:01.0421111Z 2025-09-07T08:00:01.0421216Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0421649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0422041Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0422354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0422658Z x = module(x) 2025-09-07T08:00:01.0422943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0423259Z out = self.relu(out) 2025-09-07T08:00:01.0423357Z 2025-09-07T08:00:01.0423456Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0423887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0424290Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0424607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0424910Z x = module(x) 2025-09-07T08:00:01.0425200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0425519Z out = self.relu(out) 2025-09-07T08:00:01.0425615Z 2025-09-07T08:00:01.0425723Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0426156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0426553Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0426894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0427204Z x = module(x) 2025-09-07T08:00:01.0427489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0427825Z out = self.relu(out) 2025-09-07T08:00:01.0427932Z 2025-09-07T08:00:01.0428005Z cudagraph partition due to non gpu ops 2025-09-07T08:00:01.0428231Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0428665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0429067Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0429373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0429678Z x = module(x) 2025-09-07T08:00:01.0429968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0430285Z out = self.relu(out) 2025-09-07T08:00:01.0430380Z 2025-09-07T08:00:01.0430484Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0430906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0431309Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0431624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0431930Z x = module(x) 2025-09-07T08:00:01.0432208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0432522Z out = self.relu(out) 2025-09-07T08:00:01.0432626Z 2025-09-07T08:00:01.0432721Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0433153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0433558Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0433865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0434172Z x = module(x) 2025-09-07T08:00:01.0434457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0434769Z out = self.relu(out) 2025-09-07T08:00:01.0434866Z 2025-09-07T08:00:01.0434962Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0435389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0435792Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0436106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0436413Z x = module(x) 2025-09-07T08:00:01.0436691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0437002Z out = self.relu(out) 2025-09-07T08:00:01.0437110Z 2025-09-07T08:00:01.0437187Z cudagraph partition due to non gpu ops 2025-09-07T08:00:01.0437413Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0437836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0438236Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0438548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0438862Z x = module(x) 2025-09-07T08:00:01.0439180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0439494Z out = self.relu(out) 2025-09-07T08:00:01.0439596Z 2025-09-07T08:00:01.0439691Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0440119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0440549Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0440860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0441159Z x = module(x) 2025-09-07T08:00:01.0441444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0441760Z out = self.relu(out) 2025-09-07T08:00:01.0441855Z 2025-09-07T08:00:01.0441958Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0442388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0442790Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0443100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0443409Z x = module(x) 2025-09-07T08:00:01.0443697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0444001Z out = self.relu(out) 2025-09-07T08:00:01.0444104Z 2025-09-07T08:00:01.0444202Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0444634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0445036Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0445349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0445648Z x = module(x) 2025-09-07T08:00:01.0445932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0446246Z out = self.relu(out) 2025-09-07T08:00:01.0446342Z 2025-09-07T08:00:01.0446423Z cudagraph partition due to non gpu ops 2025-09-07T08:00:01.0446645Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0447071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0447468Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0447779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0448083Z x = module(x) 2025-09-07T08:00:01.0448361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0448674Z out = self.relu(out) 2025-09-07T08:00:01.0448770Z 2025-09-07T08:00:01.0448874Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0449301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0449696Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0450007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0450311Z x = module(x) 2025-09-07T08:00:01.0450593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0450907Z out = self.relu(out) 2025-09-07T08:00:01.0451001Z 2025-09-07T08:00:01.0451099Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0451525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0451965Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0452274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0452570Z x = module(x) 2025-09-07T08:00:01.0452857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0453205Z out = self.relu(out) 2025-09-07T08:00:01.0453299Z 2025-09-07T08:00:01.0453408Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0453838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0454236Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0454550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0454859Z x = module(x) 2025-09-07T08:00:01.0455155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0455476Z out = self.relu(out) 2025-09-07T08:00:01.0455574Z 2025-09-07T08:00:01.0455649Z cudagraph partition due to non gpu ops 2025-09-07T08:00:01.0455874Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0456311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0456714Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0457020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0457330Z x = module(x) 2025-09-07T08:00:01.0457618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0457934Z out = self.relu(out) 2025-09-07T08:00:01.0458030Z 2025-09-07T08:00:01.0458135Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0458562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0458963Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0459282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0459591Z x = module(x) 2025-09-07T08:00:01.0459875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0460196Z out = self.relu(out) 2025-09-07T08:00:01.0460301Z 2025-09-07T08:00:01.0460398Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0460833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0461236Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0461544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0461850Z x = module(x) 2025-09-07T08:00:01.0462139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0462462Z out = self.relu(out) 2025-09-07T08:00:01.0462555Z 2025-09-07T08:00:01.0462660Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0463082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0463486Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0463796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0464101Z x = module(x) 2025-09-07T08:00:01.0464379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0464732Z out = self.relu(out) 2025-09-07T08:00:01.0464834Z 2025-09-07T08:00:01.0464928Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0465355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0465781Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0466087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0466391Z x = module(x) 2025-09-07T08:00:01.0466678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:00:01.0466994Z out = self.relu(out) 2025-09-07T08:00:01.0467089Z 2025-09-07T08:00:01.0467186Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0467623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0468059Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0468457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0468781Z x = module(x) 2025-09-07T08:00:01.0469079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:00:01.0469472Z out = self.relu(out) 2025-09-07T08:00:01.0469579Z 2025-09-07T08:00:01.0469656Z cudagraph partition due to non gpu ops 2025-09-07T08:00:01.0469885Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0470321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0470720Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0471123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:00:01.0471599Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0472079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:00:01.0472554Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0472735Z 2025-09-07T08:00:01.0472832Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0473270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0473669Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0473980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0474281Z x = module(x) 2025-09-07T08:00:01.0474571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0474888Z out = self.relu(out) 2025-09-07T08:00:01.0474985Z 2025-09-07T08:00:01.0475089Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0475521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0475921Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0476320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:00:01.0476779Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0477252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:00:01.0477772Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0477955Z 2025-09-07T08:00:01.0478053Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0478482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0478931Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0479247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0479557Z x = module(x) 2025-09-07T08:00:01.0479840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0480158Z out = self.relu(out) 2025-09-07T08:00:01.0480266Z 2025-09-07T08:00:01.0480363Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0480791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0481188Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0481492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0481797Z x = module(x) 2025-09-07T08:00:01.0482088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0482404Z out = self.relu(out) 2025-09-07T08:00:01.0482502Z 2025-09-07T08:00:01.0482598Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0483024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0483421Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0483733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0484040Z x = module(x) 2025-09-07T08:00:01.0484321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0484637Z out = self.relu(out) 2025-09-07T08:00:01.0484742Z 2025-09-07T08:00:01.0484817Z cudagraph partition due to non gpu ops 2025-09-07T08:00:01.0485047Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0485468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0485866Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0486176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0486485Z x = module(x) 2025-09-07T08:00:01.0486768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0487075Z out = self.relu(out) 2025-09-07T08:00:01.0487181Z 2025-09-07T08:00:01.0487276Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0487706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0488108Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0488415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0488721Z x = module(x) 2025-09-07T08:00:01.0489010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:00:01.0489328Z out = self.relu(out) 2025-09-07T08:00:01.0489425Z 2025-09-07T08:00:01.0489527Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0489947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0490376Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0490691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0490998Z x = module(x) 2025-09-07T08:00:01.0491280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0491627Z out = self.relu(out) 2025-09-07T08:00:01.0491729Z 2025-09-07T08:00:01.0491825Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0492256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:00:01.0492659Z feats = self.feat_extractor(x) 2025-09-07T08:00:01.0492965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:00:01.0493272Z x = module(x) 2025-09-07T08:00:01.0493562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:00:01.0493878Z out = self.relu(out) 2025-09-07T08:00:01.0493974Z 2025-09-07T08:00:01.0494056Z cudagraph partition due to non gpu ops 2025-09-07T08:00:01.0494273Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0494705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0495108Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0495504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:00:01.0495969Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0496437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:00:01.0496909Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0497093Z 2025-09-07T08:00:01.0497189Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0497617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0498018Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0498407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:00:01.0498868Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0499337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:00:01.0499799Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:00:01.0499980Z 2025-09-07T08:00:01.0500088Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0500512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0500917Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0501317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 86, in forward 2025-09-07T08:00:01.0501738Z out.append(self.upsample(out[-1]) + t) 2025-09-07T08:00:01.0501873Z 2025-09-07T08:00:01.0501978Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0502397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0502799Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0503225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:00:01.0503726Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0504188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:00:01.0504674Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0504850Z 2025-09-07T08:00:01.0504948Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0505379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0505779Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0506172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 86, in forward 2025-09-07T08:00:01.0506577Z out.append(self.upsample(out[-1]) + t) 2025-09-07T08:00:01.0506717Z 2025-09-07T08:00:01.0506812Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0507239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0507642Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0508037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:00:01.0508476Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0508931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:00:01.0509379Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0509542Z 2025-09-07T08:00:01.0509648Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0510080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0510471Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0510863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:00:01.0511308Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0511757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:00:01.0512208Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0512366Z 2025-09-07T08:00:01.0512463Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0512896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0513294Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0513684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:00:01.0514122Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0514559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:00:01.0515005Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0515172Z 2025-09-07T08:00:01.0515267Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0515694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0516126Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0516521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:00:01.0516972Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0517457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:00:01.0517906Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0518066Z 2025-09-07T08:00:01.0518172Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0518605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0519002Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0519399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:00:01.0519843Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0520292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:00:01.0520744Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0520907Z 2025-09-07T08:00:01.0521002Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0521431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0521833Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0522228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:00:01.0522673Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0523116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:00:01.0523568Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0523737Z 2025-09-07T08:00:01.0523833Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0524261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0524662Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0525049Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:00:01.0525488Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0525939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:00:01.0526393Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0526555Z 2025-09-07T08:00:01.0526658Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0527083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0527486Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0527878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:00:01.0528319Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0528769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:00:01.0529251Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0529414Z 2025-09-07T08:00:01.0529510Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0529939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0530366Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0530758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:00:01.0531200Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0531640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:00:01.0532091Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0532257Z 2025-09-07T08:00:01.0532356Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0532786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0533187Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0533576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:00:01.0534018Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0534464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:00:01.0534910Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0535068Z 2025-09-07T08:00:01.0535171Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0535592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:00:01.0535992Z feat_concat = self.fpn(feats) 2025-09-07T08:00:01.0536384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:00:01.0539493Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0539970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:00:01.0540428Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:00:01.0540603Z 2025-09-07T08:00:01.0540677Z cudagraph partition due to non gpu ops 2025-09-07T08:00:01.0540885Z cudagraph partition due to non gpu ops 2025-09-07T08:00:01.0541111Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0541553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:00:01.0541961Z logits = self.prob_head(feat_concat) 2025-09-07T08:00:01.0542090Z 2025-09-07T08:00:01.0542186Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0542619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:00:01.0543035Z logits = self.prob_head(feat_concat) 2025-09-07T08:00:01.0543156Z 2025-09-07T08:00:01.0543258Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0543682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:00:01.0544093Z logits = self.prob_head(feat_concat) 2025-09-07T08:00:01.0544219Z 2025-09-07T08:00:01.0544331Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0544816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:00:01.0545223Z logits = self.prob_head(feat_concat) 2025-09-07T08:00:01.0545346Z 2025-09-07T08:00:01.0545477Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0545906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:00:01.0546311Z logits = self.prob_head(feat_concat) 2025-09-07T08:00:01.0546434Z 2025-09-07T08:00:01.0546538Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:01.0546969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:00:01.0547366Z logits = self.prob_head(feat_concat) 2025-09-07T08:00:01.0547497Z 2025-09-07T08:00:01.0547574Z cudagraph partition due to non gpu ops 2025-09-07T08:00:13.8005040Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:13.8005625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/core.py", line 95, in __call__ 2025-09-07T08:00:13.8006096Z for bmap in (proba_map >= self.bin_thresh).astype(np.uint8) 2025-09-07T08:00:13.8006349Z 2025-09-07T08:00:14.3022129Z pass 2025-09-07T08:00:14.3022562Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:00:16.2816083Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:00:16.2817502Z import pynvml # type: ignore[import] 2025-09-07T08:00:17.9978896Z 2025-09-07T08:00:20.1298726Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:00:20.1299071Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:00:20.1299320Z cpu eval doctr_reco_predictor 2025-09-07T08:00:20.2973229Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:00:20.3234898Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:00:20.3459787Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:00:32.4312082Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4312593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4312983Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4313111Z 2025-09-07T08:00:32.4313227Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4313645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4314012Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4314139Z 2025-09-07T08:00:32.4314240Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4314635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4315000Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4315127Z 2025-09-07T08:00:32.4315225Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4315604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4315963Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4316082Z 2025-09-07T08:00:32.4316185Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4317521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4317989Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4318111Z 2025-09-07T08:00:32.4318210Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4318596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4319072Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4319190Z 2025-09-07T08:00:32.4319291Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4319666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4320026Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4320154Z 2025-09-07T08:00:32.4320250Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4320639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4321001Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4321119Z 2025-09-07T08:00:32.4321214Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4321601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4321962Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4322076Z 2025-09-07T08:00:32.4322178Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4322556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4322908Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4323032Z 2025-09-07T08:00:32.4323127Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4323539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4323891Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4324017Z 2025-09-07T08:00:32.4324111Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4324491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4324854Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4324970Z 2025-09-07T08:00:32.4325074Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4325446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4325805Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4325928Z 2025-09-07T08:00:32.4326020Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4326395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4326757Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4326874Z 2025-09-07T08:00:32.4326968Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4327349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4327714Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4327829Z 2025-09-07T08:00:32.4327933Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4328306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4328666Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4328791Z 2025-09-07T08:00:32.4328883Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4329262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4329655Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4329775Z 2025-09-07T08:00:32.4329869Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4330249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4330646Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4330763Z 2025-09-07T08:00:32.4330862Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4331238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4331584Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4331708Z 2025-09-07T08:00:32.4331802Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4332175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4332531Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4332651Z 2025-09-07T08:00:32.4332752Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4333123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4333479Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4333604Z 2025-09-07T08:00:32.4333700Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4334075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4334431Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4334546Z 2025-09-07T08:00:32.4334641Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4335019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4335373Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4335492Z 2025-09-07T08:00:32.4335598Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4335980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4336329Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4336457Z 2025-09-07T08:00:32.4336551Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4336926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4337283Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4337399Z 2025-09-07T08:00:32.4337494Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4337875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4338235Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4338354Z 2025-09-07T08:00:32.4338458Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4338837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4339188Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4339313Z 2025-09-07T08:00:32.4339410Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4339785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4340142Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4340258Z 2025-09-07T08:00:32.4340357Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4340726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4341080Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4341206Z 2025-09-07T08:00:32.4341330Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4341708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4342059Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4342176Z 2025-09-07T08:00:32.4342303Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:32.4342681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:00:32.4343036Z features = self.feat_extractor(x) 2025-09-07T08:00:32.4343153Z 2025-09-07T08:00:32.4343240Z cudagraph partition due to non gpu ops 2025-09-07T08:00:36.6711120Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:36.6711740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 75, in ctc_best_path 2025-09-07T08:00:36.6712255Z probs = F.softmax(logits, dim=-1).max(dim=-1).values.min(dim=1).values 2025-09-07T08:00:36.6712507Z 2025-09-07T08:00:36.8309540Z pass 2025-09-07T08:00:36.8309987Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:00:38.4847237Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:00:38.4848681Z import pynvml # type: ignore[import] 2025-09-07T08:00:40.1990319Z 2025-09-07T08:00:40.2451168Z loading model: 0it [00:00, ?it/s]Gym has been unmaintained since 2022 and does not support NumPy 2.0 amongst other critical functionality. 2025-09-07T08:00:40.2452445Z Please upgrade to Gymnasium, the maintained drop-in replacement of Gym, or contact the authors of your software and request that they upgrade. 2025-09-07T08:00:40.2453039Z Users of this version of Gym should be able to simply replace 'import gym' with 'import gymnasium as gym' in the vast majority of cases. 2025-09-07T08:00:40.2453585Z See the migration guide at https://gymnasium.farama.org/introduction/migration_guide/ for additional information. 2025-09-07T08:00:40.7311083Z 2025-09-07T08:00:40.7311645Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:00:40.7312089Z cpu eval drq 2025-09-07T08:00:40.7365114Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:00:40.7411770Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:00:40.7447211Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:00:42.9783257Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9783700Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9784063Z return mod(*inputs) 2025-09-07T08:00:42.9784324Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:00:42.9784611Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:00:42.9784755Z 2025-09-07T08:00:42.9784862Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9785218Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9785522Z return mod(*inputs) 2025-09-07T08:00:42.9785757Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:42.9786043Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:42.9786311Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:42.9786571Z h = self.forward_conv(obs) 2025-09-07T08:00:42.9786836Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 43, in forward_conv 2025-09-07T08:00:42.9787113Z conv = torch.relu(self.convs[0](obs)) 2025-09-07T08:00:42.9787628Z 2025-09-07T08:00:42.9787744Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9788084Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9788391Z return mod(*inputs) 2025-09-07T08:00:42.9788720Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:42.9788988Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:42.9789265Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:42.9789560Z h = self.forward_conv(obs) 2025-09-07T08:00:42.9789820Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 43, in forward_conv 2025-09-07T08:00:42.9790094Z conv = torch.relu(self.convs[0](obs)) 2025-09-07T08:00:42.9790220Z 2025-09-07T08:00:42.9790327Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9790662Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9790965Z return mod(*inputs) 2025-09-07T08:00:42.9791195Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:42.9791470Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:42.9791732Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:42.9791997Z h = self.forward_conv(obs) 2025-09-07T08:00:42.9792262Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:00:42.9792543Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:00:42.9792667Z 2025-09-07T08:00:42.9792772Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9793097Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9793399Z return mod(*inputs) 2025-09-07T08:00:42.9793628Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:42.9793901Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:42.9794164Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:42.9794419Z h = self.forward_conv(obs) 2025-09-07T08:00:42.9794676Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:00:42.9794952Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:00:42.9795074Z 2025-09-07T08:00:42.9795176Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9795500Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9795799Z return mod(*inputs) 2025-09-07T08:00:42.9796027Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:42.9796294Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:42.9796556Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:42.9796808Z h = self.forward_conv(obs) 2025-09-07T08:00:42.9797057Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:00:42.9797329Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:00:42.9797454Z 2025-09-07T08:00:42.9797548Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9797882Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9798179Z return mod(*inputs) 2025-09-07T08:00:42.9798406Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:42.9798673Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:42.9798930Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:42.9799177Z h = self.forward_conv(obs) 2025-09-07T08:00:42.9799427Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:00:42.9799700Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:00:42.9799880Z 2025-09-07T08:00:42.9799983Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9800314Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9800610Z return mod(*inputs) 2025-09-07T08:00:42.9800886Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:42.9801154Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:42.9801413Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:42.9801662Z h = self.forward_conv(obs) 2025-09-07T08:00:42.9801916Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:00:42.9802183Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:00:42.9802306Z 2025-09-07T08:00:42.9802402Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9802733Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9803031Z return mod(*inputs) 2025-09-07T08:00:42.9803256Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:42.9803516Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:42.9803791Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:42.9804043Z h = self.forward_conv(obs) 2025-09-07T08:00:42.9804295Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:00:42.9804563Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:00:42.9804683Z 2025-09-07T08:00:42.9804762Z cudagraph partition due to non gpu ops 2025-09-07T08:00:42.9804985Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9805314Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9805611Z return mod(*inputs) 2025-09-07T08:00:42.9805833Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:42.9806101Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:42.9806369Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 61, in forward 2025-09-07T08:00:42.9806614Z out = self.head(h) 2025-09-07T08:00:42.9806714Z 2025-09-07T08:00:42.9806819Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9807143Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9807443Z return mod(*inputs) 2025-09-07T08:00:42.9807671Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:42.9807941Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:42.9808202Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 61, in forward 2025-09-07T08:00:42.9808450Z out = self.head(h) 2025-09-07T08:00:42.9808553Z 2025-09-07T08:00:42.9808630Z cudagraph partition due to non gpu ops 2025-09-07T08:00:42.9808857Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9809213Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9809502Z return mod(*inputs) 2025-09-07T08:00:42.9809733Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:00:42.9810009Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:00:42.9810140Z 2025-09-07T08:00:42.9810245Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9810569Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9810855Z return mod(*inputs) 2025-09-07T08:00:42.9811082Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:00:42.9811349Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:00:42.9811478Z 2025-09-07T08:00:42.9811614Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9811940Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9812237Z return mod(*inputs) 2025-09-07T08:00:42.9812461Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:00:42.9812761Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:00:42.9812893Z 2025-09-07T08:00:42.9812997Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9813317Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9813610Z return mod(*inputs) 2025-09-07T08:00:42.9813837Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:00:42.9814103Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:00:42.9814234Z 2025-09-07T08:00:42.9814329Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9814655Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9814952Z return mod(*inputs) 2025-09-07T08:00:42.9815180Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:00:42.9815450Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:00:42.9815583Z 2025-09-07T08:00:42.9815678Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.9816000Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:42.9816291Z return mod(*inputs) 2025-09-07T08:00:42.9816517Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:00:42.9816773Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:00:42.9816911Z 2025-09-07T08:00:42.9816987Z cudagraph partition due to non gpu ops 2025-09-07T08:00:42.9817183Z cudagraph partition due to non gpu ops 2025-09-07T08:00:42.9817379Z cudagraph partition due to non gpu ops 2025-09-07T08:00:42.9817574Z cudagraph partition due to non gpu ops 2025-09-07T08:00:42.9817761Z cudagraph partition due to non gpu ops 2025-09-07T08:00:50.8666030Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8666458Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8666824Z return mod(*inputs) 2025-09-07T08:00:50.8667087Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:00:50.8667379Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:00:50.8667520Z 2025-09-07T08:00:50.8667631Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8667966Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8668427Z return mod(*inputs) 2025-09-07T08:00:50.8668679Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:50.8668976Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:50.8669249Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:50.8669520Z h = self.forward_conv(obs) 2025-09-07T08:00:50.8669789Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 43, in forward_conv 2025-09-07T08:00:50.8670077Z conv = torch.relu(self.convs[0](obs)) 2025-09-07T08:00:50.8670207Z 2025-09-07T08:00:50.8670317Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8670653Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8670955Z return mod(*inputs) 2025-09-07T08:00:50.8671190Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:50.8671468Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:50.8671734Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:50.8671988Z h = self.forward_conv(obs) 2025-09-07T08:00:50.8672498Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 43, in forward_conv 2025-09-07T08:00:50.8672785Z conv = torch.relu(self.convs[0](obs)) 2025-09-07T08:00:50.8672913Z 2025-09-07T08:00:50.8673012Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8673352Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8673754Z return mod(*inputs) 2025-09-07T08:00:50.8673989Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:50.8674259Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:50.8674522Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:50.8674779Z h = self.forward_conv(obs) 2025-09-07T08:00:50.8675063Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:00:50.8675338Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:00:50.8675462Z 2025-09-07T08:00:50.8675573Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8675905Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8676195Z return mod(*inputs) 2025-09-07T08:00:50.8676426Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:50.8676699Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:50.8676967Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:50.8677214Z h = self.forward_conv(obs) 2025-09-07T08:00:50.8677465Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:00:50.8677741Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:00:50.8677865Z 2025-09-07T08:00:50.8677974Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8678302Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8678609Z return mod(*inputs) 2025-09-07T08:00:50.8678840Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:50.8679112Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:50.8679378Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:50.8679627Z h = self.forward_conv(obs) 2025-09-07T08:00:50.8679882Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:00:50.8680156Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:00:50.8680280Z 2025-09-07T08:00:50.8680385Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8680707Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8681007Z return mod(*inputs) 2025-09-07T08:00:50.8681233Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:50.8681506Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:50.8681768Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:50.8682021Z h = self.forward_conv(obs) 2025-09-07T08:00:50.8682272Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:00:50.8682548Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:00:50.8682673Z 2025-09-07T08:00:50.8682775Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8683096Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8683392Z return mod(*inputs) 2025-09-07T08:00:50.8683618Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:50.8683887Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:50.8684146Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:50.8684401Z h = self.forward_conv(obs) 2025-09-07T08:00:50.8684698Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:00:50.8684980Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:00:50.8685104Z 2025-09-07T08:00:50.8685207Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8685529Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8685861Z return mod(*inputs) 2025-09-07T08:00:50.8686093Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:50.8686361Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:50.8686624Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:00:50.8686874Z h = self.forward_conv(obs) 2025-09-07T08:00:50.8687129Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:00:50.8687401Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:00:50.8687524Z 2025-09-07T08:00:50.8687611Z cudagraph partition due to non gpu ops 2025-09-07T08:00:50.8687832Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8688164Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8688458Z return mod(*inputs) 2025-09-07T08:00:50.8688687Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:50.8688946Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:50.8689214Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 61, in forward 2025-09-07T08:00:50.8689460Z out = self.head(h) 2025-09-07T08:00:50.8689556Z 2025-09-07T08:00:50.8689660Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8689977Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8690273Z return mod(*inputs) 2025-09-07T08:00:50.8690506Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:00:50.8690777Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:00:50.8691047Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 61, in forward 2025-09-07T08:00:50.8691286Z out = self.head(h) 2025-09-07T08:00:50.8691387Z 2025-09-07T08:00:50.8691467Z cudagraph partition due to non gpu ops 2025-09-07T08:00:50.8691689Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8692018Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8692307Z return mod(*inputs) 2025-09-07T08:00:50.8692538Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:00:50.8692807Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:00:50.8692938Z 2025-09-07T08:00:50.8693040Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8693367Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8693654Z return mod(*inputs) 2025-09-07T08:00:50.8693881Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:00:50.8694151Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:00:50.8694282Z 2025-09-07T08:00:50.8694384Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8694703Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8695000Z return mod(*inputs) 2025-09-07T08:00:50.8695229Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:00:50.8695497Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:00:50.8695626Z 2025-09-07T08:00:50.8695728Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8696046Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8696339Z return mod(*inputs) 2025-09-07T08:00:50.8696600Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:00:50.8696871Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:00:50.8697002Z 2025-09-07T08:00:50.8697098Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8699326Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8699624Z return mod(*inputs) 2025-09-07T08:00:50.8699854Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:00:50.8700114Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:00:50.8700257Z 2025-09-07T08:00:50.8700352Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:50.8700684Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:50.8700985Z return mod(*inputs) 2025-09-07T08:00:50.8701224Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:00:50.8701487Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:00:50.8701629Z 2025-09-07T08:00:50.8701706Z cudagraph partition due to non gpu ops 2025-09-07T08:00:50.8701913Z cudagraph partition due to non gpu ops 2025-09-07T08:00:50.8702119Z cudagraph partition due to non gpu ops 2025-09-07T08:00:50.8702315Z cudagraph partition due to non gpu ops 2025-09-07T08:00:50.8702513Z cudagraph partition due to non gpu ops 2025-09-07T08:00:51.0274611Z pass 2025-09-07T08:00:51.0278395Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:00:52.5313621Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:00:52.5315017Z import pynvml # type: ignore[import] 2025-09-07T08:00:54.2418362Z 2025-09-07T08:00:55.3378702Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:00:55.3379019Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:00:55.3379272Z cpu eval fastNLP_Bert 2025-09-07T08:00:55.8300348Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:00:56.0527276Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:00:56.2751153Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:00:56.3918868Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T08:00:56.3919505Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T08:00:56.3919987Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] or: 2025-09-07T08:00:56.3920433Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T08:00:56.3920963Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] to include these operations in the captured graph. 2025-09-07T08:00:56.3921433Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:00:56.3921850Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break: from user code at: 2025-09-07T08:00:56.3922453Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:00:56.3923037Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] return mod(*inputs) 2025-09-07T08:00:56.3923983Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/models/bert.py", line 265, in forward 2025-09-07T08:00:56.3924612Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] sequence_output = self.bert(words) 2025-09-07T08:00:56.3925362Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 137, in forward 2025-09-07T08:00:56.3925992Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] outputs = self.model(words) 2025-09-07T08:00:56.3926620Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 445, in forward 2025-09-07T08:00:56.3927799Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] max_word_piece_length = batch_word_pieces_length.sum(dim=-1).max().item() # 表示word piece的长度(包括padding) 2025-09-07T08:00:56.3930238Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:00:56.3930602Z W0907 08:00:56.390746 31707 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:01:06.0995556Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:06.0996072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 445, in forward 2025-09-07T08:01:06.0996910Z max_word_piece_length = batch_word_pieces_length.sum(dim=-1).max().item() # 表示word piece的长度(包括padding) 2025-09-07T08:01:06.0997155Z 2025-09-07T08:01:06.0997260Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:06.0997694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 441, in forward 2025-09-07T08:01:06.0998058Z seq_len = word_mask.sum(dim=-1) 2025-09-07T08:01:06.0998184Z 2025-09-07T08:01:12.3535937Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3536429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 512, in forward 2025-09-07T08:01:12.3536884Z embedding_output = self.embeddings(input_ids, token_type_ids) 2025-09-07T08:01:12.3537265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 242, in forward 2025-09-07T08:01:12.3537680Z embeddings = words_embeddings + position_embeddings + token_type_embeddings 2025-09-07T08:01:12.3537923Z 2025-09-07T08:01:12.3538006Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3538217Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3538415Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3538603Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3538809Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3539004Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3539198Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3539385Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3539576Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3539777Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3539970Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3540158Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3540385Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3540754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:01:12.3541119Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:01:12.3541470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:01:12.3542166Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:01:12.3542570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:01:12.3542956Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:01:12.3543341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:01:12.3543776Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:01:12.3544136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:01:12.3544486Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:01:12.3544632Z 2025-09-07T08:01:12.3544716Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3544915Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3545103Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3545304Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3545501Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3545695Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3545879Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3546069Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3546270Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3546462Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3546644Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3546841Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3547034Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3547224Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3547436Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3547850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:01:12.3548204Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:01:12.3548551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:01:12.3548923Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:01:12.3549292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:01:12.3549666Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:01:12.3550029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:01:12.3550389Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:01:12.3550748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:01:12.3551101Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:01:12.3551246Z 2025-09-07T08:01:12.3551329Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3551524Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3551721Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3551912Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3552104Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3552287Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3552485Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3552675Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3552877Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3553059Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3553252Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3553442Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3553633Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3553821Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3554045Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3554459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:01:12.3554813Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:01:12.3555160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:01:12.3555573Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:01:12.3555933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:01:12.3556305Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:01:12.3556671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:01:12.3557031Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:01:12.3557382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:01:12.3557728Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:01:12.3557876Z 2025-09-07T08:01:12.3557948Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3558147Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3558336Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3558537Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3558729Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3558923Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3559107Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3559298Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3559491Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3559682Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3559869Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3560061Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3560255Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3560453Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3560664Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3561020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:01:12.3561368Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:01:12.3561721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:01:12.3562087Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:01:12.3562450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:01:12.3562818Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:01:12.3563189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:01:12.3563546Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:01:12.3563903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:01:12.3564243Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:01:12.3564390Z 2025-09-07T08:01:12.3564464Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3564663Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3564860Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3565049Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3565248Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3565443Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3565633Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3565817Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3566010Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3566202Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3566395Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3566621Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3566819Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3567008Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3567229Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3567585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:01:12.3567958Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:01:12.3568390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:01:12.3568763Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:01:12.3569134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:01:12.3569513Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:01:12.3569896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:01:12.3570272Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:01:12.3570640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:01:12.3571001Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:01:12.3571146Z 2025-09-07T08:01:12.3571222Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3571427Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3571628Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3571828Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3572018Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3572219Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3572419Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3572618Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3572813Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3573010Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3573210Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3573410Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3573600Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3573801Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3574026Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3574395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:01:12.3574751Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:01:12.3575105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:01:12.3575484Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:01:12.3575904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:01:12.3576287Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:01:12.3576662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:01:12.3577028Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:01:12.3577391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:01:12.3577744Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:01:12.3577889Z 2025-09-07T08:01:12.3577973Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3578163Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3578364Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3578560Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3578760Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3578949Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3579286Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3579490Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3579695Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3579887Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3580087Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3580353Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3580550Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3580739Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3580970Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3581341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:01:12.3581704Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:01:12.3582062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:01:12.3582434Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:01:12.3582811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:01:12.3583192Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:01:12.3583572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:01:12.3583946Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:01:12.3584299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:01:12.3584651Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:01:12.3584804Z 2025-09-07T08:01:12.3584878Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3585076Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3585265Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3585461Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3585658Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3585852Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3586041Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3586239Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3586441Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3586640Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3586829Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3587028Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3587224Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3587421Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3587650Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3588014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:01:12.3588377Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:01:12.3588737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:01:12.3589117Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:01:12.3589487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:01:12.3589871Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:01:12.3590245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:01:12.3590616Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:01:12.3590976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:01:12.3591323Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:01:12.3591474Z 2025-09-07T08:01:12.3591550Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3591784Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3591989Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3592179Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3592377Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3592578Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3592810Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3592997Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3593196Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3593394Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3593590Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3593778Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3593977Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3594173Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3594400Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3594767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:01:12.3595119Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:01:12.3595473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:01:12.3595853Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:01:12.3596227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:01:12.3596600Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:01:12.3596972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:01:12.3597340Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:01:12.3597697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:01:12.3598052Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:01:12.3598197Z 2025-09-07T08:01:12.3598270Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3598470Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3598668Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3598868Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3599058Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3599257Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3599456Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3599651Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3599839Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3600042Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3600239Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3600439Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3600628Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3600827Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3601056Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3601429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:01:12.3601784Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:01:12.3602135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:01:12.3602509Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:01:12.3602882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:01:12.3603259Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:01:12.3603633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:01:12.3603995Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:01:12.3604387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:01:12.3604746Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:01:12.3604890Z 2025-09-07T08:01:12.3604997Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3605232Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3605423Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3605623Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3605818Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3606013Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3606213Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3606405Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3606599Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3606803Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3607005Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3607196Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3607395Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3607588Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3607811Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3608175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:01:12.3608532Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:01:12.3608886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:01:12.3609261Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:01:12.3609627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:01:12.3610005Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:01:12.3610380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:01:12.3610752Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:01:12.3611113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:01:12.3611463Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:01:12.3611617Z 2025-09-07T08:01:12.3611692Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3611891Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3612088Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3612276Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3612474Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3612671Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3612871Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3613074Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3613265Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3613461Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3613662Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3613858Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3614051Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3614254Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3614484Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3614857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:01:12.3615209Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:01:12.3615564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:01:12.3615940Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:01:12.3617004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:01:12.3617401Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:01:12.3617776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:01:12.3618155Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:01:12.3618567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:01:12.3618925Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:01:12.3619070Z 2025-09-07T08:01:12.3619144Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3619342Z cudagraph partition due to non gpu ops 2025-09-07T08:01:12.3619567Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3619929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 518, in forward 2025-09-07T08:01:12.3620291Z pooled_output = self.pooler(sequence_output) 2025-09-07T08:01:12.3620640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 404, in forward 2025-09-07T08:01:12.3620997Z pooled_output = self.dense(first_token_tensor) 2025-09-07T08:01:12.3621146Z 2025-09-07T08:01:12.3621250Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3621610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 518, in forward 2025-09-07T08:01:12.3621963Z pooled_output = self.pooler(sequence_output) 2025-09-07T08:01:12.3622309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 405, in forward 2025-09-07T08:01:12.3622661Z pooled_output = self.activation(pooled_output) 2025-09-07T08:01:12.3622809Z 2025-09-07T08:01:12.3622907Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:12.3623277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 518, in forward 2025-09-07T08:01:12.3623622Z pooled_output = self.pooler(sequence_output) 2025-09-07T08:01:12.3623971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 405, in forward 2025-09-07T08:01:12.3624328Z pooled_output = self.activation(pooled_output) 2025-09-07T08:01:12.3624471Z 2025-09-07T08:01:14.3597848Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:14.3598523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 140, in torch_dynamo_resume_in_forward_at_137 2025-09-07T08:01:14.3598954Z return self.dropout(outputs) 2025-09-07T08:01:14.3599309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/embedding.py", line 160, in dropout 2025-09-07T08:01:14.3599662Z return self.dropout_layer(words) 2025-09-07T08:01:14.3599787Z 2025-09-07T08:01:14.4797185Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:14.4797781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/models/bert.py", line 266, in torch_dynamo_resume_in_forward_at_265 2025-09-07T08:01:14.4798339Z logits = self.qa_outputs(sequence_output) # [batch_size, seq_len, num_labels] 2025-09-07T08:01:14.4798557Z 2025-09-07T08:01:14.4798678Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:14.4799088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/models/bert.py", line 266, in torch_dynamo_resume_in_forward_at_265 2025-09-07T08:01:14.4799529Z logits = self.qa_outputs(sequence_output) # [batch_size, seq_len, num_labels] 2025-09-07T08:01:14.4799717Z 2025-09-07T08:01:14.4799796Z cudagraph partition due to non gpu ops 2025-09-07T08:01:14.4800003Z cudagraph partition due to non gpu ops 2025-09-07T08:01:14.6759897Z pass 2025-09-07T08:01:14.6760340Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:01:16.4919812Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:01:16.4921401Z import pynvml # type: ignore[import] 2025-09-07T08:01:18.2098465Z 2025-09-07T08:01:18.5200409Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:01:18.5200803Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:01:18.5201146Z cpu eval functorch_dp_cifar10 2025-09-07T08:01:18.5434615Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:01:18.5572305Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:01:18.5681397Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:01:24.6922953Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6923386Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6923715Z return mod(*inputs) 2025-09-07T08:01:24.6924059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6924417Z return self._forward_impl(x) 2025-09-07T08:01:24.6924753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 268, in _forward_impl 2025-09-07T08:01:24.6925094Z x = self.conv1(x) 2025-09-07T08:01:24.6925214Z 2025-09-07T08:01:24.6925321Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6925681Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6925985Z return mod(*inputs) 2025-09-07T08:01:24.6926286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6926616Z return self._forward_impl(x) 2025-09-07T08:01:24.6926949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 268, in _forward_impl 2025-09-07T08:01:24.6927276Z x = self.conv1(x) 2025-09-07T08:01:24.6927371Z 2025-09-07T08:01:24.6927473Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6927807Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6928108Z return mod(*inputs) 2025-09-07T08:01:24.6928412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6928737Z return self._forward_impl(x) 2025-09-07T08:01:24.6929066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 268, in _forward_impl 2025-09-07T08:01:24.6929396Z x = self.conv1(x) 2025-09-07T08:01:24.6929615Z 2025-09-07T08:01:24.6929720Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6930053Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6930344Z return mod(*inputs) 2025-09-07T08:01:24.6930641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6930965Z return self._forward_impl(x) 2025-09-07T08:01:24.6931293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 270, in _forward_impl 2025-09-07T08:01:24.6931619Z x = self.relu(x) 2025-09-07T08:01:24.6931713Z 2025-09-07T08:01:24.6931810Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6932144Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6932438Z return mod(*inputs) 2025-09-07T08:01:24.6933057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6933386Z return self._forward_impl(x) 2025-09-07T08:01:24.6933711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 271, in _forward_impl 2025-09-07T08:01:24.6934043Z x = self.maxpool(x) 2025-09-07T08:01:24.6934239Z 2025-09-07T08:01:24.6934345Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6934673Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6934963Z return mod(*inputs) 2025-09-07T08:01:24.6935260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6935581Z return self._forward_impl(x) 2025-09-07T08:01:24.6935904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:01:24.6936223Z x = self.layer1(x) 2025-09-07T08:01:24.6936524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.6936838Z out = self.conv1(x) 2025-09-07T08:01:24.6936933Z 2025-09-07T08:01:24.6937038Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6937367Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6937663Z return mod(*inputs) 2025-09-07T08:01:24.6937960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6938276Z return self._forward_impl(x) 2025-09-07T08:01:24.6938602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:01:24.6938920Z x = self.layer1(x) 2025-09-07T08:01:24.6939213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.6939527Z out = self.conv1(x) 2025-09-07T08:01:24.6939620Z 2025-09-07T08:01:24.6939725Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6940059Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6940349Z return mod(*inputs) 2025-09-07T08:01:24.6940647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6940968Z return self._forward_impl(x) 2025-09-07T08:01:24.6941294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:01:24.6941610Z x = self.layer1(x) 2025-09-07T08:01:24.6941897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:01:24.6942210Z out = self.relu(out) 2025-09-07T08:01:24.6942313Z 2025-09-07T08:01:24.6942419Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6942752Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6943040Z return mod(*inputs) 2025-09-07T08:01:24.6943334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6943653Z return self._forward_impl(x) 2025-09-07T08:01:24.6943976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:01:24.6944294Z x = self.layer1(x) 2025-09-07T08:01:24.6944585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.6944901Z out = self.conv2(out) 2025-09-07T08:01:24.6945004Z 2025-09-07T08:01:24.6945106Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6945434Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6945771Z return mod(*inputs) 2025-09-07T08:01:24.6946070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6946391Z return self._forward_impl(x) 2025-09-07T08:01:24.6946716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:01:24.6947073Z x = self.layer1(x) 2025-09-07T08:01:24.6947365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.6947684Z out = self.conv2(out) 2025-09-07T08:01:24.6947787Z 2025-09-07T08:01:24.6947891Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6948227Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6948524Z return mod(*inputs) 2025-09-07T08:01:24.6948828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6949157Z return self._forward_impl(x) 2025-09-07T08:01:24.6949488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:01:24.6949815Z x = self.layer1(x) 2025-09-07T08:01:24.6950116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 103, in forward 2025-09-07T08:01:24.6950440Z out = self.relu(out) 2025-09-07T08:01:24.6950543Z 2025-09-07T08:01:24.6950652Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6950990Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6951287Z return mod(*inputs) 2025-09-07T08:01:24.6951590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6951917Z return self._forward_impl(x) 2025-09-07T08:01:24.6952254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:01:24.6952580Z x = self.layer1(x) 2025-09-07T08:01:24.6952877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.6953196Z out = self.conv1(x) 2025-09-07T08:01:24.6953292Z 2025-09-07T08:01:24.6953402Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6953747Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6954042Z return mod(*inputs) 2025-09-07T08:01:24.6954344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6954672Z return self._forward_impl(x) 2025-09-07T08:01:24.6955003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:01:24.6955327Z x = self.layer1(x) 2025-09-07T08:01:24.6955623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.6955939Z out = self.conv1(x) 2025-09-07T08:01:24.6956035Z 2025-09-07T08:01:24.6956143Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6956483Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6956780Z return mod(*inputs) 2025-09-07T08:01:24.6957082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6957407Z return self._forward_impl(x) 2025-09-07T08:01:24.6957735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:01:24.6958057Z x = self.layer1(x) 2025-09-07T08:01:24.6958355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:01:24.6958710Z out = self.relu(out) 2025-09-07T08:01:24.6958816Z 2025-09-07T08:01:24.6958922Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6959251Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6959576Z return mod(*inputs) 2025-09-07T08:01:24.6959874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6960200Z return self._forward_impl(x) 2025-09-07T08:01:24.6960527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:01:24.6960849Z x = self.layer1(x) 2025-09-07T08:01:24.6961151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.6961471Z out = self.conv2(out) 2025-09-07T08:01:24.6961578Z 2025-09-07T08:01:24.6961691Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6962027Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6962323Z return mod(*inputs) 2025-09-07T08:01:24.6962627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6962956Z return self._forward_impl(x) 2025-09-07T08:01:24.6963284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:01:24.6963609Z x = self.layer1(x) 2025-09-07T08:01:24.6963903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.6964222Z out = self.conv2(out) 2025-09-07T08:01:24.6964325Z 2025-09-07T08:01:24.6964433Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6964775Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6965075Z return mod(*inputs) 2025-09-07T08:01:24.6965375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6965700Z return self._forward_impl(x) 2025-09-07T08:01:24.6966035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:01:24.6966365Z x = self.layer1(x) 2025-09-07T08:01:24.6966667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 103, in forward 2025-09-07T08:01:24.6966993Z out = self.relu(out) 2025-09-07T08:01:24.6967097Z 2025-09-07T08:01:24.6967206Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6967549Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6967848Z return mod(*inputs) 2025-09-07T08:01:24.6968189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6968722Z return self._forward_impl(x) 2025-09-07T08:01:24.6969065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:01:24.6969458Z x = self.layer2(x) 2025-09-07T08:01:24.6969750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.6970071Z out = self.conv1(x) 2025-09-07T08:01:24.6970168Z 2025-09-07T08:01:24.6970277Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6970629Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6970933Z return mod(*inputs) 2025-09-07T08:01:24.6971245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6971578Z return self._forward_impl(x) 2025-09-07T08:01:24.6971991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:01:24.6972333Z x = self.layer2(x) 2025-09-07T08:01:24.6972627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.6973027Z out = self.conv1(x) 2025-09-07T08:01:24.6973123Z 2025-09-07T08:01:24.6973231Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6973580Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6973882Z return mod(*inputs) 2025-09-07T08:01:24.6974184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6974515Z return self._forward_impl(x) 2025-09-07T08:01:24.6974853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:01:24.6975196Z x = self.layer2(x) 2025-09-07T08:01:24.6975494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:01:24.6975818Z out = self.relu(out) 2025-09-07T08:01:24.6975919Z 2025-09-07T08:01:24.6976024Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6976373Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6976670Z return mod(*inputs) 2025-09-07T08:01:24.6976971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6977300Z return self._forward_impl(x) 2025-09-07T08:01:24.6977631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:01:24.6977965Z x = self.layer2(x) 2025-09-07T08:01:24.6978259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.6978581Z out = self.conv2(out) 2025-09-07T08:01:24.6978685Z 2025-09-07T08:01:24.6978792Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6979136Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6979440Z return mod(*inputs) 2025-09-07T08:01:24.6979772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6980101Z return self._forward_impl(x) 2025-09-07T08:01:24.6980438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:01:24.6980774Z x = self.layer2(x) 2025-09-07T08:01:24.6981069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.6981391Z out = self.conv2(out) 2025-09-07T08:01:24.6981493Z 2025-09-07T08:01:24.6981602Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6981949Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6982249Z return mod(*inputs) 2025-09-07T08:01:24.6982553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6982884Z return self._forward_impl(x) 2025-09-07T08:01:24.6983217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:01:24.6983551Z x = self.layer2(x) 2025-09-07T08:01:24.6983842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 100, in forward 2025-09-07T08:01:24.6984182Z identity = self.downsample(x) 2025-09-07T08:01:24.6984301Z 2025-09-07T08:01:24.6984408Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6984786Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6985085Z return mod(*inputs) 2025-09-07T08:01:24.6985389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6985716Z return self._forward_impl(x) 2025-09-07T08:01:24.6986091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:01:24.6986435Z x = self.layer2(x) 2025-09-07T08:01:24.6986737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 100, in forward 2025-09-07T08:01:24.6987078Z identity = self.downsample(x) 2025-09-07T08:01:24.6987201Z 2025-09-07T08:01:24.6987310Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6987658Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6987965Z return mod(*inputs) 2025-09-07T08:01:24.6988278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6988613Z return self._forward_impl(x) 2025-09-07T08:01:24.6988952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:01:24.6989295Z x = self.layer2(x) 2025-09-07T08:01:24.6989594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 103, in forward 2025-09-07T08:01:24.6989925Z out = self.relu(out) 2025-09-07T08:01:24.6990029Z 2025-09-07T08:01:24.6990135Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6990483Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6990784Z return mod(*inputs) 2025-09-07T08:01:24.6991092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6991426Z return self._forward_impl(x) 2025-09-07T08:01:24.6991768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:01:24.6992112Z x = self.layer2(x) 2025-09-07T08:01:24.6992406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.6992734Z out = self.conv1(x) 2025-09-07T08:01:24.6992834Z 2025-09-07T08:01:24.6992940Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6993288Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6993592Z return mod(*inputs) 2025-09-07T08:01:24.6993900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6994231Z return self._forward_impl(x) 2025-09-07T08:01:24.6994573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:01:24.6994917Z x = self.layer2(x) 2025-09-07T08:01:24.6995212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.6995539Z out = self.conv1(x) 2025-09-07T08:01:24.6995642Z 2025-09-07T08:01:24.6995752Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6996104Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6996410Z return mod(*inputs) 2025-09-07T08:01:24.6996719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6997054Z return self._forward_impl(x) 2025-09-07T08:01:24.6997394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:01:24.6997741Z x = self.layer2(x) 2025-09-07T08:01:24.6998068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:01:24.6998392Z out = self.relu(out) 2025-09-07T08:01:24.6998493Z 2025-09-07T08:01:24.6998603Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.6998978Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.6999276Z return mod(*inputs) 2025-09-07T08:01:24.6999578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.6999903Z return self._forward_impl(x) 2025-09-07T08:01:24.7000232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:01:24.7000572Z x = self.layer2(x) 2025-09-07T08:01:24.7000861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.7001186Z out = self.conv2(out) 2025-09-07T08:01:24.7001291Z 2025-09-07T08:01:24.7001399Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7001742Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7002041Z return mod(*inputs) 2025-09-07T08:01:24.7002353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7002679Z return self._forward_impl(x) 2025-09-07T08:01:24.7003016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:01:24.7003350Z x = self.layer2(x) 2025-09-07T08:01:24.7003638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.7003960Z out = self.conv2(out) 2025-09-07T08:01:24.7004062Z 2025-09-07T08:01:24.7004168Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7004511Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7004810Z return mod(*inputs) 2025-09-07T08:01:24.7005110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7005438Z return self._forward_impl(x) 2025-09-07T08:01:24.7005773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:01:24.7006107Z x = self.layer2(x) 2025-09-07T08:01:24.7006399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 103, in forward 2025-09-07T08:01:24.7006721Z out = self.relu(out) 2025-09-07T08:01:24.7006821Z 2025-09-07T08:01:24.7006927Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7007264Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7007561Z return mod(*inputs) 2025-09-07T08:01:24.7007868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7008192Z return self._forward_impl(x) 2025-09-07T08:01:24.7008524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:01:24.7008860Z x = self.layer3(x) 2025-09-07T08:01:24.7009146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.7009462Z out = self.conv1(x) 2025-09-07T08:01:24.7009559Z 2025-09-07T08:01:24.7009666Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7010007Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7010305Z return mod(*inputs) 2025-09-07T08:01:24.7010639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7010969Z return self._forward_impl(x) 2025-09-07T08:01:24.7011307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:01:24.7011645Z x = self.layer3(x) 2025-09-07T08:01:24.7013521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.7013846Z out = self.conv1(x) 2025-09-07T08:01:24.7013947Z 2025-09-07T08:01:24.7014055Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7014394Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7014691Z return mod(*inputs) 2025-09-07T08:01:24.7014994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7015327Z return self._forward_impl(x) 2025-09-07T08:01:24.7015665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:01:24.7016003Z x = self.layer3(x) 2025-09-07T08:01:24.7016296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:01:24.7016625Z out = self.relu(out) 2025-09-07T08:01:24.7016725Z 2025-09-07T08:01:24.7016832Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7017176Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7017470Z return mod(*inputs) 2025-09-07T08:01:24.7017769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7018093Z return self._forward_impl(x) 2025-09-07T08:01:24.7018426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:01:24.7018762Z x = self.layer3(x) 2025-09-07T08:01:24.7019060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.7019371Z out = self.conv2(out) 2025-09-07T08:01:24.7019470Z 2025-09-07T08:01:24.7019573Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7019912Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7020205Z return mod(*inputs) 2025-09-07T08:01:24.7020508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7020832Z return self._forward_impl(x) 2025-09-07T08:01:24.7021163Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:01:24.7021500Z x = self.layer3(x) 2025-09-07T08:01:24.7021791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.7022111Z out = self.conv2(out) 2025-09-07T08:01:24.7022211Z 2025-09-07T08:01:24.7022316Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7022652Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7022950Z return mod(*inputs) 2025-09-07T08:01:24.7023254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7023630Z return self._forward_impl(x) 2025-09-07T08:01:24.7023953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:01:24.7024283Z x = self.layer3(x) 2025-09-07T08:01:24.7024569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 100, in forward 2025-09-07T08:01:24.7024889Z identity = self.downsample(x) 2025-09-07T08:01:24.7025005Z 2025-09-07T08:01:24.7025168Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7025504Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7025790Z return mod(*inputs) 2025-09-07T08:01:24.7026090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7026444Z return self._forward_impl(x) 2025-09-07T08:01:24.7026770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:01:24.7027116Z x = self.layer3(x) 2025-09-07T08:01:24.7027401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 100, in forward 2025-09-07T08:01:24.7027722Z identity = self.downsample(x) 2025-09-07T08:01:24.7027844Z 2025-09-07T08:01:24.7027940Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7028274Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7028563Z return mod(*inputs) 2025-09-07T08:01:24.7028859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7029178Z return self._forward_impl(x) 2025-09-07T08:01:24.7029503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:01:24.7029826Z x = self.layer3(x) 2025-09-07T08:01:24.7030110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 103, in forward 2025-09-07T08:01:24.7030425Z out = self.relu(out) 2025-09-07T08:01:24.7030531Z 2025-09-07T08:01:24.7030627Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7030958Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7031244Z return mod(*inputs) 2025-09-07T08:01:24.7031540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7031858Z return self._forward_impl(x) 2025-09-07T08:01:24.7032184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:01:24.7032516Z x = self.layer3(x) 2025-09-07T08:01:24.7032801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.7033109Z out = self.conv1(x) 2025-09-07T08:01:24.7033210Z 2025-09-07T08:01:24.7033306Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7033635Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7033921Z return mod(*inputs) 2025-09-07T08:01:24.7034212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7034530Z return self._forward_impl(x) 2025-09-07T08:01:24.7034852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:01:24.7035172Z x = self.layer3(x) 2025-09-07T08:01:24.7035450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.7035761Z out = self.conv1(x) 2025-09-07T08:01:24.7035858Z 2025-09-07T08:01:24.7035952Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7036280Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7036569Z return mod(*inputs) 2025-09-07T08:01:24.7036860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7037180Z return self._forward_impl(x) 2025-09-07T08:01:24.7037534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:01:24.7037859Z x = self.layer3(x) 2025-09-07T08:01:24.7038138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:01:24.7038450Z out = self.relu(out) 2025-09-07T08:01:24.7038585Z 2025-09-07T08:01:24.7038684Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7039018Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7039308Z return mod(*inputs) 2025-09-07T08:01:24.7039609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7039934Z return self._forward_impl(x) 2025-09-07T08:01:24.7040260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:01:24.7040590Z x = self.layer3(x) 2025-09-07T08:01:24.7040878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.7041196Z out = self.conv2(out) 2025-09-07T08:01:24.7041306Z 2025-09-07T08:01:24.7041406Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7041738Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7042035Z return mod(*inputs) 2025-09-07T08:01:24.7042330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7042652Z return self._forward_impl(x) 2025-09-07T08:01:24.7042976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:01:24.7043301Z x = self.layer3(x) 2025-09-07T08:01:24.7043587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.7043905Z out = self.conv2(out) 2025-09-07T08:01:24.7044015Z 2025-09-07T08:01:24.7044113Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7044448Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7044739Z return mod(*inputs) 2025-09-07T08:01:24.7045041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7045363Z return self._forward_impl(x) 2025-09-07T08:01:24.7045688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:01:24.7046014Z x = self.layer3(x) 2025-09-07T08:01:24.7046301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 103, in forward 2025-09-07T08:01:24.7046614Z out = self.relu(out) 2025-09-07T08:01:24.7046718Z 2025-09-07T08:01:24.7046820Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7047157Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7047446Z return mod(*inputs) 2025-09-07T08:01:24.7047744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7048070Z return self._forward_impl(x) 2025-09-07T08:01:24.7048398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:01:24.7048728Z x = self.layer4(x) 2025-09-07T08:01:24.7049009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.7049320Z out = self.conv1(x) 2025-09-07T08:01:24.7049420Z 2025-09-07T08:01:24.7049514Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7071391Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7071735Z return mod(*inputs) 2025-09-07T08:01:24.7072247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7072602Z return self._forward_impl(x) 2025-09-07T08:01:24.7072951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:01:24.7073394Z x = self.layer4(x) 2025-09-07T08:01:24.7073705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.7074023Z out = self.conv1(x) 2025-09-07T08:01:24.7074136Z 2025-09-07T08:01:24.7074245Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7074597Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7074906Z return mod(*inputs) 2025-09-07T08:01:24.7075207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7075545Z return self._forward_impl(x) 2025-09-07T08:01:24.7075883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:01:24.7076220Z x = self.layer4(x) 2025-09-07T08:01:24.7076520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:01:24.7076838Z out = self.relu(out) 2025-09-07T08:01:24.7076947Z 2025-09-07T08:01:24.7077048Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7077397Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7077700Z return mod(*inputs) 2025-09-07T08:01:24.7077993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7078318Z return self._forward_impl(x) 2025-09-07T08:01:24.7078653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:01:24.7078983Z x = self.layer4(x) 2025-09-07T08:01:24.7079275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.7079585Z out = self.conv2(out) 2025-09-07T08:01:24.7079700Z 2025-09-07T08:01:24.7079799Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7080141Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7080446Z return mod(*inputs) 2025-09-07T08:01:24.7080741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7081066Z return self._forward_impl(x) 2025-09-07T08:01:24.7081395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:01:24.7081724Z x = self.layer4(x) 2025-09-07T08:01:24.7082021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.7082332Z out = self.conv2(out) 2025-09-07T08:01:24.7082442Z 2025-09-07T08:01:24.7082539Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7082881Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7083184Z return mod(*inputs) 2025-09-07T08:01:24.7083478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7083800Z return self._forward_impl(x) 2025-09-07T08:01:24.7084129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:01:24.7084463Z x = self.layer4(x) 2025-09-07T08:01:24.7084758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 100, in forward 2025-09-07T08:01:24.7085114Z identity = self.downsample(x) 2025-09-07T08:01:24.7085243Z 2025-09-07T08:01:24.7085342Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7085678Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7086007Z return mod(*inputs) 2025-09-07T08:01:24.7086294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7086613Z return self._forward_impl(x) 2025-09-07T08:01:24.7086938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:01:24.7087265Z x = self.layer4(x) 2025-09-07T08:01:24.7087562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 100, in forward 2025-09-07T08:01:24.7087885Z identity = self.downsample(x) 2025-09-07T08:01:24.7088011Z 2025-09-07T08:01:24.7088112Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7088451Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7088752Z return mod(*inputs) 2025-09-07T08:01:24.7089045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7089371Z return self._forward_impl(x) 2025-09-07T08:01:24.7089700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:01:24.7090028Z x = self.layer4(x) 2025-09-07T08:01:24.7090325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 103, in forward 2025-09-07T08:01:24.7090636Z out = self.relu(out) 2025-09-07T08:01:24.7090749Z 2025-09-07T08:01:24.7090848Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7091188Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7091489Z return mod(*inputs) 2025-09-07T08:01:24.7091780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7092097Z return self._forward_impl(x) 2025-09-07T08:01:24.7092432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:01:24.7092763Z x = self.layer4(x) 2025-09-07T08:01:24.7093056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.7093361Z out = self.conv1(x) 2025-09-07T08:01:24.7093463Z 2025-09-07T08:01:24.7093560Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7093895Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7094195Z return mod(*inputs) 2025-09-07T08:01:24.7094484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7094805Z return self._forward_impl(x) 2025-09-07T08:01:24.7095130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:01:24.7095460Z x = self.layer4(x) 2025-09-07T08:01:24.7095753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:01:24.7096057Z out = self.conv1(x) 2025-09-07T08:01:24.7096161Z 2025-09-07T08:01:24.7096256Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7096588Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7096888Z return mod(*inputs) 2025-09-07T08:01:24.7097177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7097495Z return self._forward_impl(x) 2025-09-07T08:01:24.7097896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:01:24.7098231Z x = self.layer4(x) 2025-09-07T08:01:24.7098528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:01:24.7098869Z out = self.relu(out) 2025-09-07T08:01:24.7098979Z 2025-09-07T08:01:24.7099076Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7099412Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7099715Z return mod(*inputs) 2025-09-07T08:01:24.7100005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7100328Z return self._forward_impl(x) 2025-09-07T08:01:24.7100654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:01:24.7100986Z x = self.layer4(x) 2025-09-07T08:01:24.7101278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.7101584Z out = self.conv2(out) 2025-09-07T08:01:24.7101693Z 2025-09-07T08:01:24.7101795Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7102132Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7102433Z return mod(*inputs) 2025-09-07T08:01:24.7102721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7103042Z return self._forward_impl(x) 2025-09-07T08:01:24.7103363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:01:24.7103695Z x = self.layer4(x) 2025-09-07T08:01:24.7103985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:01:24.7104294Z out = self.conv2(out) 2025-09-07T08:01:24.7104404Z 2025-09-07T08:01:24.7104500Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:24.7104832Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:24.7105141Z return mod(*inputs) 2025-09-07T08:01:24.7105440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:01:24.7105757Z return self._forward_impl(x) 2025-09-07T08:01:24.7106087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 278, in _forward_impl 2025-09-07T08:01:24.7106418Z x = self.avgpool(x) 2025-09-07T08:01:24.7106510Z 2025-09-07T08:01:24.7106598Z cudagraph partition due to non gpu ops 2025-09-07T08:01:24.7106801Z cudagraph partition due to non gpu ops 2025-09-07T08:01:41.7187935Z pass 2025-09-07T08:01:41.7188433Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:01:43.3115341Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:01:43.3116936Z import pynvml # type: ignore[import] 2025-09-07T08:01:45.0258326Z 2025-09-07T08:01:45.1529576Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:01:45.1529968Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:01:45.1530305Z cpu eval functorch_maml_omniglot 2025-09-07T08:01:45.1566283Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:01:45.1605377Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:01:45.1631200Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:01:47.4632468Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:47.4632925Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:47.4633717Z return mod(*inputs) 2025-09-07T08:01:47.4633827Z 2025-09-07T08:01:47.4633933Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:47.4634285Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:47.4634593Z return mod(*inputs) 2025-09-07T08:01:47.4634690Z 2025-09-07T08:01:47.4634799Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:47.4635139Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:47.4635432Z return mod(*inputs) 2025-09-07T08:01:47.4635535Z 2025-09-07T08:01:47.4635643Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:47.4635979Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:47.4636277Z return mod(*inputs) 2025-09-07T08:01:47.4636375Z 2025-09-07T08:01:47.4636479Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:47.4636844Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:47.4637169Z return mod(*inputs) 2025-09-07T08:01:47.4637262Z 2025-09-07T08:01:47.4637367Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:47.4637756Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:47.4638066Z return mod(*inputs) 2025-09-07T08:01:47.4638177Z 2025-09-07T08:01:47.4638273Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:47.4638628Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:47.4638958Z return mod(*inputs) 2025-09-07T08:01:47.4639053Z 2025-09-07T08:01:47.4639173Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:47.4639521Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:47.4639848Z return mod(*inputs) 2025-09-07T08:01:47.4639950Z 2025-09-07T08:01:47.4640047Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:47.4640401Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:47.4640722Z return mod(*inputs) 2025-09-07T08:01:47.4640816Z 2025-09-07T08:01:47.4640911Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:47.4641274Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:47.4641571Z return mod(*inputs) 2025-09-07T08:01:47.4641663Z 2025-09-07T08:01:47.4641764Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:47.4642088Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:47.4642390Z return mod(*inputs) 2025-09-07T08:01:47.4642490Z 2025-09-07T08:01:47.4642584Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:47.4642918Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:47.4643215Z return mod(*inputs) 2025-09-07T08:01:47.4643306Z 2025-09-07T08:01:47.4643399Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:47.4643733Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:47.4644031Z return mod(*inputs) 2025-09-07T08:01:47.4644123Z 2025-09-07T08:01:47.4644222Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:01:47.4644542Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:01:47.4644980Z return mod(*inputs) 2025-09-07T08:01:47.4645091Z 2025-09-07T08:01:47.4645175Z cudagraph partition due to non gpu ops 2025-09-07T08:01:47.4645386Z cudagraph partition due to non gpu ops 2025-09-07T08:01:55.5474058Z pass 2025-09-07T08:01:55.5474474Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:01:57.0224998Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:01:57.0226388Z import pynvml # type: ignore[import] 2025-09-07T08:01:58.7389068Z 2025-09-07T08:02:00.0165011Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:02:00.0165337Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:02:00.0165593Z cpu eval hf_Albert 2025-09-07T08:02:00.4768944Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:02:00.5614247Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:02:00.6438160Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:02:09.7407256Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7407700Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7408035Z return mod(**inputs) 2025-09-07T08:02:09.7408428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7408802Z outputs = self.albert( 2025-09-07T08:02:09.7409162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 700, in forward 2025-09-07T08:02:09.7409585Z embedding_output = self.embeddings( 2025-09-07T08:02:09.7409972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 239, in forward 2025-09-07T08:02:09.7410357Z embeddings += position_embeddings 2025-09-07T08:02:09.7410483Z 2025-09-07T08:02:09.7410579Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7410785Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7410985Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7411185Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7411376Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7411577Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7411777Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7411978Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7412170Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7412367Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7412566Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7412774Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7412968Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7413168Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7413371Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7413571Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7413767Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7413964Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7414162Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7414416Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7414619Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7414849Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7415207Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7415515Z return mod(**inputs) 2025-09-07T08:02:09.7416254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7416641Z outputs = self.albert( 2025-09-07T08:02:09.7417012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7417475Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7417856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7418274Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7418689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7419186Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7419678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:02:09.7420183Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:02:09.7420666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:02:09.7421112Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:09.7421303Z 2025-09-07T08:02:09.7421387Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7421582Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7421786Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7421984Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7422215Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7422556Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7422867Z return mod(**inputs) 2025-09-07T08:02:09.7423225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7423598Z outputs = self.albert( 2025-09-07T08:02:09.7423960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7424330Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7424702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7425110Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7425514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7426005Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7426492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:02:09.7426901Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:02:09.7427313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:02:09.7427731Z return forward_fn(*input_tensors) 2025-09-07T08:02:09.7428120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:02:09.7428543Z ffn_output = self.activation(ffn_output) 2025-09-07T08:02:09.7428888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:02:09.7429331Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:02:09.7429557Z 2025-09-07T08:02:09.7429646Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7429897Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7430102Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7430303Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7430503Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7430701Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7430933Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7431129Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7431327Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7431557Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7431897Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7432211Z return mod(**inputs) 2025-09-07T08:02:09.7432570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7432942Z outputs = self.albert( 2025-09-07T08:02:09.7433295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7433671Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7434039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7434456Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7434887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7435370Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7435861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:02:09.7436341Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:02:09.7436821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:02:09.7437259Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:09.7437445Z 2025-09-07T08:02:09.7437521Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7437729Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7437934Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7438131Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7438350Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7438702Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7439008Z return mod(**inputs) 2025-09-07T08:02:09.7439366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7439737Z outputs = self.albert( 2025-09-07T08:02:09.7440088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7440467Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7440834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7441249Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7441651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7442141Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7442632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:02:09.7443016Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:02:09.7443434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:02:09.7443810Z return forward_fn(*input_tensors) 2025-09-07T08:02:09.7444193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:02:09.7444610Z ffn_output = self.activation(ffn_output) 2025-09-07T08:02:09.7444951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:02:09.7445377Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:02:09.7445594Z 2025-09-07T08:02:09.7445671Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7445876Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7446074Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7446272Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7446461Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7446659Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7446852Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7447050Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7447246Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7447471Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7447820Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7448130Z return mod(**inputs) 2025-09-07T08:02:09.7448481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7448845Z outputs = self.albert( 2025-09-07T08:02:09.7449203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7449576Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7449949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7450364Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7450763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7451254Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7451739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:02:09.7452215Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:02:09.7452694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:02:09.7453126Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:09.7453321Z 2025-09-07T08:02:09.7453397Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7453594Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7453796Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7453986Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7454221Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7454565Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7454873Z return mod(**inputs) 2025-09-07T08:02:09.7455219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7455590Z outputs = self.albert( 2025-09-07T08:02:09.7455942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7456345Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7456719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7457129Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7457543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7459703Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7460203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:02:09.7460591Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:02:09.7460980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:02:09.7461368Z return forward_fn(*input_tensors) 2025-09-07T08:02:09.7461751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:02:09.7462148Z ffn_output = self.activation(ffn_output) 2025-09-07T08:02:09.7462491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:02:09.7462917Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:02:09.7463147Z 2025-09-07T08:02:09.7463222Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7463428Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7463631Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7463826Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7464029Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7464226Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7464425Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7464619Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7464822Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7465054Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7465404Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7465721Z return mod(**inputs) 2025-09-07T08:02:09.7466068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7466437Z outputs = self.albert( 2025-09-07T08:02:09.7466794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7467170Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7467532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7467949Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7468455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7468950Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7469447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:02:09.7469925Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:02:09.7470409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:02:09.7470855Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:09.7471052Z 2025-09-07T08:02:09.7471132Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7471423Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7471618Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7471821Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7472049Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7472397Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7472750Z return mod(**inputs) 2025-09-07T08:02:09.7473103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7473477Z outputs = self.albert( 2025-09-07T08:02:09.7473832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7474207Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7474568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7474981Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7475391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7475886Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7476375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:02:09.7476750Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:02:09.7477135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:02:09.7477514Z return forward_fn(*input_tensors) 2025-09-07T08:02:09.7477886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:02:09.7478273Z ffn_output = self.activation(ffn_output) 2025-09-07T08:02:09.7478607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:02:09.7479040Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:02:09.7479272Z 2025-09-07T08:02:09.7479347Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7479550Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7479742Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7480021Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7480220Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7480417Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7480616Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7480806Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7481004Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7481237Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7481584Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7481890Z return mod(**inputs) 2025-09-07T08:02:09.7482250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7482626Z outputs = self.albert( 2025-09-07T08:02:09.7482982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7483350Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7483720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7484132Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7484542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7485061Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7485548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:02:09.7486065Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:02:09.7486544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:02:09.7486982Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:09.7487168Z 2025-09-07T08:02:09.7487252Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7487446Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7487643Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7487840Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7488065Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7488404Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7488717Z return mod(**inputs) 2025-09-07T08:02:09.7489069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7489440Z outputs = self.albert( 2025-09-07T08:02:09.7489790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7490155Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7490523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7490933Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7491338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7491822Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7492303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:02:09.7492686Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:02:09.7493070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:02:09.7493444Z return forward_fn(*input_tensors) 2025-09-07T08:02:09.7493817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:02:09.7494203Z ffn_output = self.activation(ffn_output) 2025-09-07T08:02:09.7494546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:02:09.7494974Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:02:09.7495191Z 2025-09-07T08:02:09.7495274Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7495468Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7495672Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7495874Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7496073Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7496262Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7496461Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7496656Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7496853Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7497081Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7497419Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7497731Z return mod(**inputs) 2025-09-07T08:02:09.7498122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7498503Z outputs = self.albert( 2025-09-07T08:02:09.7498855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7499257Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7499627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7500038Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7500447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7500925Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7501416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:02:09.7501898Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:02:09.7502374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:02:09.7502815Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:09.7503001Z 2025-09-07T08:02:09.7503075Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7503278Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7503477Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7503674Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7503893Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7504239Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7504552Z return mod(**inputs) 2025-09-07T08:02:09.7504900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7505266Z outputs = self.albert( 2025-09-07T08:02:09.7505611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7505990Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7506356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7506763Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7507162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7507648Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7508142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:02:09.7508522Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:02:09.7508901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:02:09.7509272Z return forward_fn(*input_tensors) 2025-09-07T08:02:09.7509655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:02:09.7510041Z ffn_output = self.activation(ffn_output) 2025-09-07T08:02:09.7510377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:02:09.7510802Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:02:09.7511023Z 2025-09-07T08:02:09.7511132Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7511341Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7511543Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7511742Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7511934Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7512172Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7512371Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7512572Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7512763Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7512994Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7513346Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7513672Z return mod(**inputs) 2025-09-07T08:02:09.7514022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7514378Z outputs = self.albert( 2025-09-07T08:02:09.7514727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7515095Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7515453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7515851Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7516253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7516729Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7517210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:02:09.7517681Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:02:09.7518153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:02:09.7518577Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:09.7518771Z 2025-09-07T08:02:09.7518843Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7519043Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7519241Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7519425Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7519647Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7519984Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7520288Z return mod(**inputs) 2025-09-07T08:02:09.7520626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7520989Z outputs = self.albert( 2025-09-07T08:02:09.7521334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7521695Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7522058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7522453Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7522849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7523325Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7523803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:02:09.7524205Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:02:09.7524574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:02:09.7524938Z return forward_fn(*input_tensors) 2025-09-07T08:02:09.7525306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:02:09.7525713Z ffn_output = self.activation(ffn_output) 2025-09-07T08:02:09.7526043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:02:09.7526456Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:02:09.7526677Z 2025-09-07T08:02:09.7526751Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7526952Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7527151Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7527338Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7527535Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7527729Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7527921Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7528108Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7528307Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7528533Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7528873Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7529178Z return mod(**inputs) 2025-09-07T08:02:09.7529517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7529876Z outputs = self.albert( 2025-09-07T08:02:09.7530222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7530590Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7530944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7531347Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7531745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7532223Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7532700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:02:09.7533158Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:02:09.7533631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:02:09.7534057Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:09.7534237Z 2025-09-07T08:02:09.7534318Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7534515Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7534701Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7534896Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7535116Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7535451Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7535744Z return mod(**inputs) 2025-09-07T08:02:09.7536084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7536440Z outputs = self.albert( 2025-09-07T08:02:09.7536814Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7537177Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7537531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7537929Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7538360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7538837Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7539311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:02:09.7539675Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:02:09.7540052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:02:09.7540422Z return forward_fn(*input_tensors) 2025-09-07T08:02:09.7540790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:02:09.7541171Z ffn_output = self.activation(ffn_output) 2025-09-07T08:02:09.7541500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:02:09.7541918Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:02:09.7542144Z 2025-09-07T08:02:09.7542218Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7542420Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7542611Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7542804Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7542997Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7543193Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7543380Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7543575Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7543772Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7544000Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7544334Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7544641Z return mod(**inputs) 2025-09-07T08:02:09.7544989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7545355Z outputs = self.albert( 2025-09-07T08:02:09.7545704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7546063Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7546429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7546833Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7547235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7547713Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7548187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:02:09.7548658Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:02:09.7549127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:02:09.7549555Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:09.7549738Z 2025-09-07T08:02:09.7549852Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7550044Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7550242Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7550437Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7550661Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7551021Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7551326Z return mod(**inputs) 2025-09-07T08:02:09.7551671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7552030Z outputs = self.albert( 2025-09-07T08:02:09.7552373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7552724Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7553088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7553485Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7553878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7554347Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7554823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:02:09.7555196Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:02:09.7555567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:02:09.7555934Z return forward_fn(*input_tensors) 2025-09-07T08:02:09.7556297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:02:09.7556679Z ffn_output = self.activation(ffn_output) 2025-09-07T08:02:09.7557011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:02:09.7557428Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:02:09.7557650Z 2025-09-07T08:02:09.7557733Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7557922Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7558119Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7558317Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7558511Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7558696Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7558889Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7559084Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7559284Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7559500Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7559838Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7560142Z return mod(**inputs) 2025-09-07T08:02:09.7560485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7560851Z outputs = self.albert( 2025-09-07T08:02:09.7561191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7561552Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7561920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7562318Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7562758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7563239Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7563713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:02:09.7564213Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:02:09.7564682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:02:09.7565109Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:09.7565292Z 2025-09-07T08:02:09.7565363Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7565560Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7565756Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7565953Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7566168Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7566500Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7566797Z return mod(**inputs) 2025-09-07T08:02:09.7567140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7567492Z outputs = self.albert( 2025-09-07T08:02:09.7567833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7568194Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7568612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7569013Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7569409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7569892Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7570373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:02:09.7570754Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:02:09.7571135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:02:09.7571502Z return forward_fn(*input_tensors) 2025-09-07T08:02:09.7571876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:02:09.7572266Z ffn_output = self.activation(ffn_output) 2025-09-07T08:02:09.7572604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:02:09.7573024Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:02:09.7573240Z 2025-09-07T08:02:09.7573314Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7573518Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7573715Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7573910Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7574101Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7574297Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7574493Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7574689Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7574878Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7575104Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7575509Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7575813Z return mod(**inputs) 2025-09-07T08:02:09.7576149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7576509Z outputs = self.albert( 2025-09-07T08:02:09.7576927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7577293Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7577656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7578050Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7578449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7578928Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7579404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:02:09.7579873Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:02:09.7580337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:02:09.7580766Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:09.7580958Z 2025-09-07T08:02:09.7581033Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7581234Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7581431Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7581618Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7581837Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7582172Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7582471Z return mod(**inputs) 2025-09-07T08:02:09.7582802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7583165Z outputs = self.albert( 2025-09-07T08:02:09.7583507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7583870Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7584227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7584614Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7585009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7585486Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7585959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:02:09.7586329Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:02:09.7586698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:02:09.7587068Z return forward_fn(*input_tensors) 2025-09-07T08:02:09.7587432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:02:09.7587808Z ffn_output = self.activation(ffn_output) 2025-09-07T08:02:09.7588131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:02:09.7588602Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:02:09.7588824Z 2025-09-07T08:02:09.7588899Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7589104Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7589309Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7589503Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7589733Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7589926Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7590123Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7590308Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7590503Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7590724Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7591064Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7591360Z return mod(**inputs) 2025-09-07T08:02:09.7591707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7592066Z outputs = self.albert( 2025-09-07T08:02:09.7592411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7592774Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7593131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7593536Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7593939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7594417Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7594895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:02:09.7595359Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:02:09.7595832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:02:09.7596263Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:09.7596445Z 2025-09-07T08:02:09.7596528Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7596721Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7596908Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7597100Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7597319Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:09.7597656Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:09.7597949Z return mod(**inputs) 2025-09-07T08:02:09.7598293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:02:09.7598658Z outputs = self.albert( 2025-09-07T08:02:09.7599003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:02:09.7599370Z encoder_outputs = self.encoder( 2025-09-07T08:02:09.7599723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:02:09.7600121Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:02:09.7600513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:02:09.7600988Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:02:09.7601522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:02:09.7601891Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:02:09.7602267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:02:09.7602636Z return forward_fn(*input_tensors) 2025-09-07T08:02:09.7603035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:02:09.7603415Z ffn_output = self.activation(ffn_output) 2025-09-07T08:02:09.7603754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:02:09.7604177Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:02:09.7604399Z 2025-09-07T08:02:09.7604487Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7604689Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7604884Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7605086Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7605287Z cudagraph partition due to non gpu ops 2025-09-07T08:02:09.7605485Z cudagraph partition due to non gpu ops 2025-09-07T08:02:21.2099935Z pass 2025-09-07T08:02:21.2100482Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:02:22.9421282Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:02:22.9422700Z import pynvml # type: ignore[import] 2025-09-07T08:02:24.6562999Z 2025-09-07T08:02:27.9546712Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:02:27.9547025Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:02:27.9547309Z cpu eval hf_Bart 2025-09-07T08:02:28.6407918Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:02:28.8197474Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:02:28.9752727Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:02:41.1776636Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1777001Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1777212Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1777406Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1777606Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1777828Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1778023Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1778207Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1778399Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1778627Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1778820Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1779004Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1779209Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1779418Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1779630Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1779837Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1780073Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1780464Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1780817Z return mod(**inputs) 2025-09-07T08:02:41.1781215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1781676Z outputs = self.model( 2025-09-07T08:02:41.1782295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:02:41.1782695Z encoder_outputs = self.encoder( 2025-09-07T08:02:41.1783105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:02:41.1783594Z layer_outputs = encoder_layer( 2025-09-07T08:02:41.1783972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1784349Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1784750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:02:41.1785156Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:02:41.1785570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1785986Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1786428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.1786913Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.1787110Z 2025-09-07T08:02:41.1787216Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1787592Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1787919Z return mod(**inputs) 2025-09-07T08:02:41.1788278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1788688Z outputs = self.model( 2025-09-07T08:02:41.1789059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:02:41.1789449Z encoder_outputs = self.encoder( 2025-09-07T08:02:41.1789821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:02:41.1790208Z layer_outputs = encoder_layer( 2025-09-07T08:02:41.1790565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1790946Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1791322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:02:41.1791766Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:02:41.1792147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1792535Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1792957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.1793392Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.1793547Z 2025-09-07T08:02:41.1793623Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1793824Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1794028Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1794228Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1794418Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1794611Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1794806Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1795009Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1795196Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1795392Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1795587Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1795779Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1795965Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1796209Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1796406Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1796599Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1796811Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1797188Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1797486Z return mod(**inputs) 2025-09-07T08:02:41.1797832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1798189Z outputs = self.model( 2025-09-07T08:02:41.1798515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:02:41.1798869Z encoder_outputs = self.encoder( 2025-09-07T08:02:41.1799224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:02:41.1799582Z layer_outputs = encoder_layer( 2025-09-07T08:02:41.1799900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1800236Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1800599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:02:41.1800981Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:02:41.1801348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1801717Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1802133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.1802584Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.1802756Z 2025-09-07T08:02:41.1802863Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1803197Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1803491Z return mod(**inputs) 2025-09-07T08:02:41.1803826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1804180Z outputs = self.model( 2025-09-07T08:02:41.1804525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:02:41.1804879Z encoder_outputs = self.encoder( 2025-09-07T08:02:41.1805223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:02:41.1805575Z layer_outputs = encoder_layer( 2025-09-07T08:02:41.1805906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1806244Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1806596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:02:41.1806970Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:02:41.1807335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1807713Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1808123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.1808541Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.1808697Z 2025-09-07T08:02:41.1808772Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1810939Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1811186Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1811375Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1811576Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1811774Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1812017Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1812204Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1812396Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1812590Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1812783Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1812975Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1813161Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1813351Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1813543Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1813731Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1813950Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1814283Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1814590Z return mod(**inputs) 2025-09-07T08:02:41.1814928Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1815279Z outputs = self.model( 2025-09-07T08:02:41.1815619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:02:41.1815977Z encoder_outputs = self.encoder( 2025-09-07T08:02:41.1816326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:02:41.1816679Z layer_outputs = encoder_layer( 2025-09-07T08:02:41.1816999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1817337Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1817696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:02:41.1818067Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:02:41.1818425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1818805Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1819223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.1819668Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.1819842Z 2025-09-07T08:02:41.1819948Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1820280Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1820594Z return mod(**inputs) 2025-09-07T08:02:41.1820930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1821284Z outputs = self.model( 2025-09-07T08:02:41.1821623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:02:41.1821978Z encoder_outputs = self.encoder( 2025-09-07T08:02:41.1822328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:02:41.1822678Z layer_outputs = encoder_layer( 2025-09-07T08:02:41.1823007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1823335Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1823728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:02:41.1824102Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:02:41.1824471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1824845Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1825306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.1825740Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.1825902Z 2025-09-07T08:02:41.1825980Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1826189Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1826395Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1826590Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1826792Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1826995Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1827193Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1827384Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1827587Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1827783Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1827985Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1828174Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1828372Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1828571Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1828765Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1828957Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1829187Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1829533Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1829845Z return mod(**inputs) 2025-09-07T08:02:41.1830184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1830545Z outputs = self.model( 2025-09-07T08:02:41.1830891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:02:41.1831262Z encoder_outputs = self.encoder( 2025-09-07T08:02:41.1831622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:02:41.1831977Z layer_outputs = encoder_layer( 2025-09-07T08:02:41.1832314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1832665Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1833033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:02:41.1833401Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:02:41.1833790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1834176Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1834595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.1835054Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.1835228Z 2025-09-07T08:02:41.1835330Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1835674Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1835982Z return mod(**inputs) 2025-09-07T08:02:41.1836328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1836690Z outputs = self.model( 2025-09-07T08:02:41.1837065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:02:41.1837431Z encoder_outputs = self.encoder( 2025-09-07T08:02:41.1837784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:02:41.1838171Z layer_outputs = encoder_layer( 2025-09-07T08:02:41.1838493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1838831Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1839194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:02:41.1839566Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:02:41.1839933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1840304Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1840714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.1841141Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.1841295Z 2025-09-07T08:02:41.1841379Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1841580Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1841768Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1841960Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1842154Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1842343Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1842527Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1842718Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1842912Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1843105Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1843290Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1843479Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1843678Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1843872Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1844061Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1844254Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1844474Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1844811Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1845108Z return mod(**inputs) 2025-09-07T08:02:41.1845448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1845801Z outputs = self.model( 2025-09-07T08:02:41.1846140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:02:41.1846501Z encoder_outputs = self.encoder( 2025-09-07T08:02:41.1846849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:02:41.1847205Z layer_outputs = encoder_layer( 2025-09-07T08:02:41.1847540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1847881Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1848232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:02:41.1848604Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:02:41.1848977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1849359Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1849801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.1850243Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.1850421Z 2025-09-07T08:02:41.1850522Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1850890Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1851195Z return mod(**inputs) 2025-09-07T08:02:41.1851535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1851884Z outputs = self.model( 2025-09-07T08:02:41.1852222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:02:41.1852583Z encoder_outputs = self.encoder( 2025-09-07T08:02:41.1852938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:02:41.1853290Z layer_outputs = encoder_layer( 2025-09-07T08:02:41.1853625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1853972Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1854334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:02:41.1854705Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:02:41.1855066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1855439Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1855852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.1856281Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.1856429Z 2025-09-07T08:02:41.1856513Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1856706Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1856900Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1857097Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1857288Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1857473Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1857669Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1857864Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1858058Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1858243Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1858438Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1858634Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1858947Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1859139Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1859342Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1859541Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1859772Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1860116Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1860436Z return mod(**inputs) 2025-09-07T08:02:41.1860785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1861161Z outputs = self.model( 2025-09-07T08:02:41.1861515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:02:41.1861883Z encoder_outputs = self.encoder( 2025-09-07T08:02:41.1862244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:02:41.1862645Z layer_outputs = encoder_layer( 2025-09-07T08:02:41.1862989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1863337Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1863706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:02:41.1864122Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:02:41.1864497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1864877Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1865294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.1865752Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.1865935Z 2025-09-07T08:02:41.1866039Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1866384Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1866700Z return mod(**inputs) 2025-09-07T08:02:41.1867041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1867408Z outputs = self.model( 2025-09-07T08:02:41.1867754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:02:41.1868118Z encoder_outputs = self.encoder( 2025-09-07T08:02:41.1868539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:02:41.1868901Z layer_outputs = encoder_layer( 2025-09-07T08:02:41.1869244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1869598Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1869969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:02:41.1870343Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:02:41.1870736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1871124Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1871551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.1871986Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.1872141Z 2025-09-07T08:02:41.1872219Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1872429Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1872637Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1872836Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1873025Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1873220Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1873419Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1873617Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1873808Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1874005Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1874199Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1874395Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1874583Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1874784Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1874985Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1875184Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1875402Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1875832Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1876145Z return mod(**inputs) 2025-09-07T08:02:41.1876492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1876907Z outputs = self.model( 2025-09-07T08:02:41.1877246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1877619Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1877990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1878359Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1878691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1879044Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1879418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:02:41.1879813Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:02:41.1880204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1880589Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1881016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.1881479Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.1881652Z 2025-09-07T08:02:41.1881763Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1882112Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1882414Z return mod(**inputs) 2025-09-07T08:02:41.1882761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1883122Z outputs = self.model( 2025-09-07T08:02:41.1883468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1883839Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1884193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1884557Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1884893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1885239Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1885601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:02:41.1885995Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:02:41.1886383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1886771Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1887198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.1887628Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.1887788Z 2025-09-07T08:02:41.1887864Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1888068Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1888270Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1888462Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1888662Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1888859Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1889089Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1889283Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1889485Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1889685Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1889887Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1890106Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1890341Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1890685Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1891001Z return mod(**inputs) 2025-09-07T08:02:41.1891347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1891700Z outputs = self.model( 2025-09-07T08:02:41.1892045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1892413Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1892771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1893124Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1893464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1893803Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1894169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:02:41.1894564Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:02:41.1894952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1895338Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1895763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.1896217Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.1896389Z 2025-09-07T08:02:41.1896495Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1896834Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1897145Z return mod(**inputs) 2025-09-07T08:02:41.1897488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1897851Z outputs = self.model( 2025-09-07T08:02:41.1898189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1898554Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1898914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1899280Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1899612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1899948Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1900316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:02:41.1900708Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:02:41.1901100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1901481Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1901893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.1902357Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.1902522Z 2025-09-07T08:02:41.1902598Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1902805Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1902998Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1903722Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1903926Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1904127Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1904328Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1904520Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1904721Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1904923Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1905125Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1905318Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1905523Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1905727Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1905925Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1906117Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1906343Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1906684Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1906995Z return mod(**inputs) 2025-09-07T08:02:41.1907329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1907691Z outputs = self.model( 2025-09-07T08:02:41.1908036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1908403Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1908760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1909124Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1909460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1909809Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1910178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:02:41.1910564Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:02:41.1910952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1911333Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1911753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.1912210Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.1912382Z 2025-09-07T08:02:41.1912485Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1912829Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1913136Z return mod(**inputs) 2025-09-07T08:02:41.1913473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1913839Z outputs = self.model( 2025-09-07T08:02:41.1914174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1914539Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1914898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1915256Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1915616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1915970Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1916342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:02:41.1916733Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:02:41.1917145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1917521Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1917944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.1918381Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.1918534Z 2025-09-07T08:02:41.1918619Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1918822Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1919020Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1919220Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1919417Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1919613Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1919801Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1920006Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1920207Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1920401Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1920590Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1920785Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1921013Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1921360Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1921665Z return mod(**inputs) 2025-09-07T08:02:41.1922014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1922376Z outputs = self.model( 2025-09-07T08:02:41.1922721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1923080Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1923451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1923817Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1924154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1924506Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1924866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:02:41.1925267Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:02:41.1925666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1926053Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1926479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.1926929Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.1927116Z 2025-09-07T08:02:41.1927215Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1927559Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1927869Z return mod(**inputs) 2025-09-07T08:02:41.1928217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1928574Z outputs = self.model( 2025-09-07T08:02:41.1928958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1929329Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1929698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1930092Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1930426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1930774Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1931144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:02:41.1931541Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:02:41.1931927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1932322Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1932754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.1933189Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.1933345Z 2025-09-07T08:02:41.1933430Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1933626Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1933833Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1934035Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1934233Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1934426Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1934624Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1934826Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1935025Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1935213Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1935414Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1935615Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1935814Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1936006Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1936202Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1936403Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1936632Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1936969Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1937278Z return mod(**inputs) 2025-09-07T08:02:41.1937623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1937989Z outputs = self.model( 2025-09-07T08:02:41.1938326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1938700Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1939062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1939427Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1939767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1940112Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1940483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:02:41.1940875Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:02:41.1941264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1941648Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1942096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.1942558Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.1942745Z 2025-09-07T08:02:41.1942848Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1943217Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1943525Z return mod(**inputs) 2025-09-07T08:02:41.1943860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1944220Z outputs = self.model( 2025-09-07T08:02:41.1944563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1944927Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1945282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1945650Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1945982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1946330Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1946703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:02:41.1947083Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:02:41.1947471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1947858Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1948282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.1948718Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.1948871Z 2025-09-07T08:02:41.1948949Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1949152Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1949354Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1949568Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1949761Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1949958Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1950153Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1950351Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1950539Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1950737Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1950929Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1951126Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1951344Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1951694Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1952006Z return mod(**inputs) 2025-09-07T08:02:41.1952353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1952719Z outputs = self.model( 2025-09-07T08:02:41.1953057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1953422Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1953781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1954145Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1954476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1954824Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1955221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:02:41.1955613Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:02:41.1956005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1956415Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1956835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.1957288Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.1957459Z 2025-09-07T08:02:41.1957566Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1957910Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1958215Z return mod(**inputs) 2025-09-07T08:02:41.1958563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1958927Z outputs = self.model( 2025-09-07T08:02:41.1959271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1959633Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1959993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1960357Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1960689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1961039Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1961401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:02:41.1961802Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:02:41.1962194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1962580Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1963008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.1963437Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.1963596Z 2025-09-07T08:02:41.1963671Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1963877Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1964081Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1964273Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1964476Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1964674Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1964873Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1965063Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1965261Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1965455Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1965651Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1965845Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1966046Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1966243Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1966439Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1966627Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1966854Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1967204Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1967516Z return mod(**inputs) 2025-09-07T08:02:41.1967903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1968307Z outputs = self.model( 2025-09-07T08:02:41.1968660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1969028Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1969520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1969883Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1970222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1970576Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1970953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:02:41.1971341Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:02:41.1971722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1972114Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1972537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.1972997Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.1973171Z 2025-09-07T08:02:41.1973281Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1973616Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1973928Z return mod(**inputs) 2025-09-07T08:02:41.1974284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1974635Z outputs = self.model( 2025-09-07T08:02:41.1974964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1975325Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1975673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1976031Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1976359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1976689Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1977051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:02:41.1977430Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:02:41.1977810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1978185Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1978590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.1979015Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.1979173Z 2025-09-07T08:02:41.1979247Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1979447Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1979636Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1979830Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1980023Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1980220Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1980402Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1980597Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1980788Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1980981Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1981204Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1981405Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1981628Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1981967Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1982298Z return mod(**inputs) 2025-09-07T08:02:41.1982629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1982983Z outputs = self.model( 2025-09-07T08:02:41.1983320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1983677Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1984022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1984382Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1984709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1985047Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1985403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:02:41.1985786Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:02:41.1986170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1986547Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1986957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.1987400Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.1987570Z 2025-09-07T08:02:41.1987671Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1988007Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1988308Z return mod(**inputs) 2025-09-07T08:02:41.1988643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1988993Z outputs = self.model( 2025-09-07T08:02:41.1989329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1989687Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1990043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1990398Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1990719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.1991061Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.1991422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:02:41.1991809Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:02:41.1992198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.1992570Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.1992985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.1993411Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.1993559Z 2025-09-07T08:02:41.1993641Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1993833Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1994067Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1994269Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1994463Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1994649Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1994844Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1995065Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1995257Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1995437Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1995631Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1995823Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1996018Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1996209Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1996397Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1996587Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.1996809Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.1997148Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.1997444Z return mod(**inputs) 2025-09-07T08:02:41.1997783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.1998144Z outputs = self.model( 2025-09-07T08:02:41.1998478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.1998832Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.1999180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.1999536Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.1999863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.2000200Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.2000549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:02:41.2000928Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:02:41.2001303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.2001679Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.2002084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.2002521Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.2002695Z 2025-09-07T08:02:41.2002795Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.2003133Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.2003437Z return mod(**inputs) 2025-09-07T08:02:41.2003767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.2004122Z outputs = self.model( 2025-09-07T08:02:41.2004458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.2004819Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.2005173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.2005526Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.2005855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.2006193Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.2006551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:02:41.2006972Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:02:41.2007344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.2007721Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.2008165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.2008594Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.2008743Z 2025-09-07T08:02:41.2008819Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2009022Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2009223Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2009423Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2009619Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2009806Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2010003Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2010200Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2010393Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2010581Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2010774Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2010971Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2011196Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.2011534Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.2011844Z return mod(**inputs) 2025-09-07T08:02:41.2012184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.2012540Z outputs = self.model( 2025-09-07T08:02:41.2012870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.2013231Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.2013588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.2013948Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.2014277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.2014613Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.2014974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:02:41.2015363Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:02:41.2015748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.2016126Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.2016533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.2016979Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.2017154Z 2025-09-07T08:02:41.2017252Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.2017589Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.2017883Z return mod(**inputs) 2025-09-07T08:02:41.2018228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.2018579Z outputs = self.model( 2025-09-07T08:02:41.2018919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.2019278Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.2019652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.2020011Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.2020338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.2020678Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.2021073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:02:41.2021461Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:02:41.2021843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.2022221Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.2022633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.2023050Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.2023206Z 2025-09-07T08:02:41.2023282Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2023481Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2023676Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2023872Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2024062Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2024255Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2024449Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2024643Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2024828Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2025021Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2025211Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2025406Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2025590Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2025784Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2025979Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2026171Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2026385Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.2026720Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.2027026Z return mod(**inputs) 2025-09-07T08:02:41.2027359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.2027704Z outputs = self.model( 2025-09-07T08:02:41.2028038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.2028391Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.2028743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.2029097Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.2029419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.2029754Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.2030110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:02:41.2030494Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:02:41.2030872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.2031237Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.2031646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.2032088Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.2032256Z 2025-09-07T08:02:41.2032389Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.2032725Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.2033028Z return mod(**inputs) 2025-09-07T08:02:41.2033362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.2033751Z outputs = self.model( 2025-09-07T08:02:41.2034098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.2034453Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.2034813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.2035175Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.2035508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.2035851Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.2036210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:02:41.2036595Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:02:41.2036979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.2037361Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.2037769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.2038200Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.2038360Z 2025-09-07T08:02:41.2038441Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2038649Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2038851Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2039049Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2039249Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2039448Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2039649Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2039842Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2040041Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2040241Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2040437Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2040629Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2040856Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.2041197Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.2041504Z return mod(**inputs) 2025-09-07T08:02:41.2041839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.2042200Z outputs = self.model( 2025-09-07T08:02:41.2042541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.2042901Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.2043266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.2043617Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.2043954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.2044297Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.2044662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:02:41.2045051Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:02:41.2045462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.2045839Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.2046252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:02:41.2046732Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:02:41.2046901Z 2025-09-07T08:02:41.2046998Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.2047338Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.2047641Z return mod(**inputs) 2025-09-07T08:02:41.2047984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:02:41.2048338Z outputs = self.model( 2025-09-07T08:02:41.2048674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:02:41.2049035Z decoder_outputs = self.decoder( 2025-09-07T08:02:41.2049389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:02:41.2049754Z layer_outputs = decoder_layer( 2025-09-07T08:02:41.2050078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:02:41.2050413Z return super().__call__(*args, **kwargs) 2025-09-07T08:02:41.2050771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:02:41.2051153Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:02:41.2051536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:02:41.2051907Z attn_output, attn_weights = attention_interface( 2025-09-07T08:02:41.2052322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:02:41.2052746Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:02:41.2052895Z 2025-09-07T08:02:41.2052975Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2053174Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2053366Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2053559Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2053751Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2053945Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2054133Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2054326Z cudagraph partition due to non gpu ops 2025-09-07T08:02:41.2054546Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:02:41.2054883Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:02:41.2055181Z return mod(**inputs) 2025-09-07T08:02:41.2055517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1491, in forward 2025-09-07T08:02:41.2055926Z lm_logits = lm_logits + self.final_logits_bias.to(lm_logits.device) 2025-09-07T08:02:41.2056099Z 2025-09-07T08:02:55.9728002Z pass 2025-09-07T08:02:55.9729339Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:02:57.9640605Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:02:57.9642011Z import pynvml # type: ignore[import] 2025-09-07T08:02:59.6860301Z 2025-09-07T08:03:02.0647599Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:03:02.0648332Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:03:02.0649715Z cpu eval hf_Bert 2025-09-07T08:03:02.5835769Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:03:02.6950085Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:03:02.8051609Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:03:12.4995479Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:12.4995931Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:12.4996252Z return mod(**inputs) 2025-09-07T08:03:12.4996727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:12.4997283Z outputs = self.bert( 2025-09-07T08:03:12.4997671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 964, in forward 2025-09-07T08:03:12.4998096Z embedding_output = self.embeddings( 2025-09-07T08:03:12.4998622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 187, in forward 2025-09-07T08:03:12.4998994Z embeddings += position_embeddings 2025-09-07T08:03:12.4999134Z 2025-09-07T08:03:12.4999214Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.4999451Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.4999742Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5000000Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5000265Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5000518Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5000785Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5001042Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5001312Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5001566Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5001831Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5002094Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5002354Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5002613Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5002877Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5003135Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5003395Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5003641Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5003898Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5004198Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:12.5004679Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:12.5005102Z return mod(**inputs) 2025-09-07T08:03:12.5005588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:12.5006084Z outputs = self.bert( 2025-09-07T08:03:12.5006560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:12.5007092Z encoder_outputs = self.encoder( 2025-09-07T08:03:12.5007592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:12.5008096Z layer_outputs = layer_module( 2025-09-07T08:03:12.5008553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:12.5009036Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:12.5009557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:12.5010086Z self_attention_outputs = self.attention( 2025-09-07T08:03:12.5010978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5011490Z return func(*args, **kwargs) 2025-09-07T08:03:12.5011979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:12.5012577Z self_outputs = self.self( 2025-09-07T08:03:12.5013060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5013567Z return func(*args, **kwargs) 2025-09-07T08:03:12.5014055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:12.5014635Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:12.5014871Z 2025-09-07T08:03:12.5014970Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5015242Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5015459Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5015659Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5015843Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5016036Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5016234Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5016427Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5016611Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5016805Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5017001Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5017194Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5017380Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5017604Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:12.5017948Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:12.5018260Z return mod(**inputs) 2025-09-07T08:03:12.5018593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:12.5018948Z outputs = self.bert( 2025-09-07T08:03:12.5019287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:12.5019658Z encoder_outputs = self.encoder( 2025-09-07T08:03:12.5020018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:12.5020368Z layer_outputs = layer_module( 2025-09-07T08:03:12.5020700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:12.5021045Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:12.5021409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:12.5021778Z self_attention_outputs = self.attention( 2025-09-07T08:03:12.5022132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5022483Z return func(*args, **kwargs) 2025-09-07T08:03:12.5022836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:12.5023192Z self_outputs = self.self( 2025-09-07T08:03:12.5023527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5023879Z return func(*args, **kwargs) 2025-09-07T08:03:12.5024229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:12.5024643Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:12.5024813Z 2025-09-07T08:03:12.5024997Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5025195Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5025396Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5025595Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5025790Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5026014Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5026211Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5026408Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5026606Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5026792Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5026984Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5027178Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5027372Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5027586Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:12.5027926Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:12.5028233Z return mod(**inputs) 2025-09-07T08:03:12.5028574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:12.5028925Z outputs = self.bert( 2025-09-07T08:03:12.5029254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:12.5029610Z encoder_outputs = self.encoder( 2025-09-07T08:03:12.5029961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:12.5030314Z layer_outputs = layer_module( 2025-09-07T08:03:12.5030706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:12.5031046Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:12.5031410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:12.5031776Z self_attention_outputs = self.attention( 2025-09-07T08:03:12.5032133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5032476Z return func(*args, **kwargs) 2025-09-07T08:03:12.5032822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:12.5033175Z self_outputs = self.self( 2025-09-07T08:03:12.5033520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5033872Z return func(*args, **kwargs) 2025-09-07T08:03:12.5034208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:12.5034618Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:12.5034795Z 2025-09-07T08:03:12.5034868Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5035068Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5035260Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5035459Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5035657Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5035854Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5036041Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5036238Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5036429Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5036621Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5036807Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5037001Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5037191Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5037409Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:12.5037775Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:12.5038084Z return mod(**inputs) 2025-09-07T08:03:12.5038423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:12.5038810Z outputs = self.bert( 2025-09-07T08:03:12.5039144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:12.5039494Z encoder_outputs = self.encoder( 2025-09-07T08:03:12.5039844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:12.5040195Z layer_outputs = layer_module( 2025-09-07T08:03:12.5040522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:12.5040864Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:12.5041213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:12.5041579Z self_attention_outputs = self.attention( 2025-09-07T08:03:12.5041938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5042289Z return func(*args, **kwargs) 2025-09-07T08:03:12.5042625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:12.5042979Z self_outputs = self.self( 2025-09-07T08:03:12.5043315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5043660Z return func(*args, **kwargs) 2025-09-07T08:03:12.5044005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:12.5044409Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:12.5044587Z 2025-09-07T08:03:12.5044662Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5044864Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5045063Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5045251Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5045444Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5045636Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5045831Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5046018Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5046237Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5046427Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5046621Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5046804Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5046996Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5047222Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:12.5047560Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:12.5047865Z return mod(**inputs) 2025-09-07T08:03:12.5048191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:12.5048547Z outputs = self.bert( 2025-09-07T08:03:12.5048883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:12.5049239Z encoder_outputs = self.encoder( 2025-09-07T08:03:12.5049583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:12.5049940Z layer_outputs = layer_module( 2025-09-07T08:03:12.5050294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:12.5050631Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:12.5050996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:12.5051354Z self_attention_outputs = self.attention( 2025-09-07T08:03:12.5051755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5052101Z return func(*args, **kwargs) 2025-09-07T08:03:12.5052448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:12.5052791Z self_outputs = self.self( 2025-09-07T08:03:12.5053124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5053465Z return func(*args, **kwargs) 2025-09-07T08:03:12.5053809Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:12.5054213Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:12.5054383Z 2025-09-07T08:03:12.5054455Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5054657Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5054852Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5055048Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5055236Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5055427Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5055620Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5055814Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5055998Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5056191Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5056383Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5056574Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5056761Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5056984Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:12.5057321Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:12.5057624Z return mod(**inputs) 2025-09-07T08:03:12.5057960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:12.5058305Z outputs = self.bert( 2025-09-07T08:03:12.5058637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:12.5058994Z encoder_outputs = self.encoder( 2025-09-07T08:03:12.5059344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:12.5059688Z layer_outputs = layer_module( 2025-09-07T08:03:12.5060015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:12.5060351Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:12.5060706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:12.5061069Z self_attention_outputs = self.attention( 2025-09-07T08:03:12.5061416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5061763Z return func(*args, **kwargs) 2025-09-07T08:03:12.5062107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:12.5062459Z self_outputs = self.self( 2025-09-07T08:03:12.5062790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5063135Z return func(*args, **kwargs) 2025-09-07T08:03:12.5063508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:12.5063915Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:12.5064084Z 2025-09-07T08:03:12.5066121Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5066320Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5066521Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5066721Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5066921Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5067104Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5067295Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5067486Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5067678Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5067860Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5068054Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5068464Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5068675Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5068900Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:12.5069251Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:12.5069601Z return mod(**inputs) 2025-09-07T08:03:12.5069953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:12.5070315Z outputs = self.bert( 2025-09-07T08:03:12.5070654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:12.5071024Z encoder_outputs = self.encoder( 2025-09-07T08:03:12.5071388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:12.5071760Z layer_outputs = layer_module( 2025-09-07T08:03:12.5072084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:12.5072430Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:12.5072800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:12.5073179Z self_attention_outputs = self.attention( 2025-09-07T08:03:12.5073544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5073890Z return func(*args, **kwargs) 2025-09-07T08:03:12.5074245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:12.5074609Z self_outputs = self.self( 2025-09-07T08:03:12.5074957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5075313Z return func(*args, **kwargs) 2025-09-07T08:03:12.5075659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:12.5076073Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:12.5076253Z 2025-09-07T08:03:12.5076327Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5076533Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5076727Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5076924Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5077117Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5077313Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5077503Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5077700Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5077897Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5078166Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5078362Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5078564Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5078767Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5079000Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:12.5079400Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:12.5079703Z return mod(**inputs) 2025-09-07T08:03:12.5080045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:12.5080401Z outputs = self.bert( 2025-09-07T08:03:12.5080744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:12.5081102Z encoder_outputs = self.encoder( 2025-09-07T08:03:12.5081462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:12.5081824Z layer_outputs = layer_module( 2025-09-07T08:03:12.5082157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:12.5082502Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:12.5082862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:12.5083232Z self_attention_outputs = self.attention( 2025-09-07T08:03:12.5083595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5083948Z return func(*args, **kwargs) 2025-09-07T08:03:12.5084294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:12.5084650Z self_outputs = self.self( 2025-09-07T08:03:12.5084998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5085346Z return func(*args, **kwargs) 2025-09-07T08:03:12.5085692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:12.5086101Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:12.5086281Z 2025-09-07T08:03:12.5086357Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5086559Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5086756Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5086950Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5087147Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5087342Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5087538Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5087726Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5087924Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5088120Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5088313Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5088498Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5088694Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5088921Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:12.5089262Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:12.5089570Z return mod(**inputs) 2025-09-07T08:03:12.5089905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:12.5090265Z outputs = self.bert( 2025-09-07T08:03:12.5090607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:12.5090980Z encoder_outputs = self.encoder( 2025-09-07T08:03:12.5091364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:12.5091727Z layer_outputs = layer_module( 2025-09-07T08:03:12.5092062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:12.5092438Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:12.5092803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:12.5093162Z self_attention_outputs = self.attention( 2025-09-07T08:03:12.5093522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5093899Z return func(*args, **kwargs) 2025-09-07T08:03:12.5094251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:12.5094607Z self_outputs = self.self( 2025-09-07T08:03:12.5094949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5095299Z return func(*args, **kwargs) 2025-09-07T08:03:12.5095652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:12.5096070Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:12.5096241Z 2025-09-07T08:03:12.5096314Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5096515Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5096713Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5096907Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5097094Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5097290Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5097486Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5097685Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5097876Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5098071Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5098269Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5098468Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5098660Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5098884Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:12.5099227Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:12.5099540Z return mod(**inputs) 2025-09-07T08:03:12.5099882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:12.5100235Z outputs = self.bert( 2025-09-07T08:03:12.5100577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:12.5100947Z encoder_outputs = self.encoder( 2025-09-07T08:03:12.5101308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:12.5101661Z layer_outputs = layer_module( 2025-09-07T08:03:12.5102002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:12.5102346Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:12.5102708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:12.5103077Z self_attention_outputs = self.attention( 2025-09-07T08:03:12.5103433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5103789Z return func(*args, **kwargs) 2025-09-07T08:03:12.5104188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:12.5104555Z self_outputs = self.self( 2025-09-07T08:03:12.5104901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5105250Z return func(*args, **kwargs) 2025-09-07T08:03:12.5105633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:12.5106047Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:12.5106222Z 2025-09-07T08:03:12.5106303Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5106499Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5106703Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5106902Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5107098Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5107284Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5107485Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5107681Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5107878Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5108064Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5108258Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5108457Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5108651Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5108874Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:12.5109209Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:12.5109515Z return mod(**inputs) 2025-09-07T08:03:12.5109859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:12.5110220Z outputs = self.bert( 2025-09-07T08:03:12.5110555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:12.5110921Z encoder_outputs = self.encoder( 2025-09-07T08:03:12.5111279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:12.5111644Z layer_outputs = layer_module( 2025-09-07T08:03:12.5111984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:12.5112326Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:12.5112693Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:12.5113065Z self_attention_outputs = self.attention( 2025-09-07T08:03:12.5113430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5113780Z return func(*args, **kwargs) 2025-09-07T08:03:12.5114135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:12.5114494Z self_outputs = self.self( 2025-09-07T08:03:12.5114837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5115193Z return func(*args, **kwargs) 2025-09-07T08:03:12.5115535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:12.5115948Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:12.5116127Z 2025-09-07T08:03:12.5116202Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5116405Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5116599Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5116796Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5116991Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5117280Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5117473Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5117672Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5117867Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5118065Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5118291Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5118491Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5118689Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5118916Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:12.5119258Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:12.5119561Z return mod(**inputs) 2025-09-07T08:03:12.5119902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:12.5120263Z outputs = self.bert( 2025-09-07T08:03:12.5120604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:12.5120962Z encoder_outputs = self.encoder( 2025-09-07T08:03:12.5121318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:12.5121684Z layer_outputs = layer_module( 2025-09-07T08:03:12.5122015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:12.5122359Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:12.5122720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:12.5123091Z self_attention_outputs = self.attention( 2025-09-07T08:03:12.5123456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5123814Z return func(*args, **kwargs) 2025-09-07T08:03:12.5124159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:12.5124520Z self_outputs = self.self( 2025-09-07T08:03:12.5124864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:12.5125222Z return func(*args, **kwargs) 2025-09-07T08:03:12.5125574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:12.5125981Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:12.5126159Z 2025-09-07T08:03:12.5126233Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5126430Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5126633Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5126823Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5127023Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5127223Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5127418Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5127603Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5127802Z cudagraph partition due to non gpu ops 2025-09-07T08:03:12.5128000Z cudagraph partition due to non gpu ops 2025-09-07T08:03:23.7098115Z pass 2025-09-07T08:03:23.7098539Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:03:25.5322845Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:03:25.5324344Z import pynvml # type: ignore[import] 2025-09-07T08:03:27.2579381Z 2025-09-07T08:03:31.1198963Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:03:31.1199297Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:03:31.1199551Z cpu eval hf_Bert_large 2025-09-07T08:03:32.4057996Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:03:32.6638397Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:03:32.9196493Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:03:49.9661772Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9662225Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9662547Z return mod(**inputs) 2025-09-07T08:03:49.9662937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9663311Z outputs = self.bert( 2025-09-07T08:03:49.9663692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 964, in forward 2025-09-07T08:03:49.9664062Z embedding_output = self.embeddings( 2025-09-07T08:03:49.9664431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 187, in forward 2025-09-07T08:03:49.9664809Z embeddings += position_embeddings 2025-09-07T08:03:49.9664937Z 2025-09-07T08:03:49.9665025Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9665226Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9665414Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9665609Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9665804Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9666015Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9666211Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9666403Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9666604Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9666797Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9666980Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9667173Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9667365Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9667561Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9667744Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9667937Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9668129Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9668400Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9668588Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9668781Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9668976Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9669171Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9669355Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9669553Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9669746Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9669941Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9670125Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9670332Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9670530Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9670721Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9670958Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9671184Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9671531Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9671830Z return mod(**inputs) 2025-09-07T08:03:49.9672173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9672532Z outputs = self.bert( 2025-09-07T08:03:49.9673140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9673513Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9673877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9674332Z layer_outputs = layer_module( 2025-09-07T08:03:49.9674680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9675025Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9675388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9675758Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9676128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9676487Z return func(*args, **kwargs) 2025-09-07T08:03:49.9676875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9677233Z self_outputs = self.self( 2025-09-07T08:03:49.9677576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9677929Z return func(*args, **kwargs) 2025-09-07T08:03:49.9678279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9678687Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9678871Z 2025-09-07T08:03:49.9678946Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9679156Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9679352Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9679540Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9679739Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9679934Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9680126Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9680310Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9680502Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9680698Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9680896Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9681082Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9681289Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9681514Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9681857Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9682163Z return mod(**inputs) 2025-09-07T08:03:49.9682498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9682859Z outputs = self.bert( 2025-09-07T08:03:49.9683196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9683561Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9683912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9684268Z layer_outputs = layer_module( 2025-09-07T08:03:49.9684602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9684948Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9685312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9685672Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9686083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9686431Z return func(*args, **kwargs) 2025-09-07T08:03:49.9686777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9687126Z self_outputs = self.self( 2025-09-07T08:03:49.9687510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9687868Z return func(*args, **kwargs) 2025-09-07T08:03:49.9688219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9688635Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9688807Z 2025-09-07T08:03:49.9688887Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9689095Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9689296Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9689505Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9689699Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9689904Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9690102Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9690301Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9690502Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9690693Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9690893Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9691093Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9691294Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9691518Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9691862Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9692173Z return mod(**inputs) 2025-09-07T08:03:49.9692526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9692888Z outputs = self.bert( 2025-09-07T08:03:49.9693227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9693595Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9693950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9694308Z layer_outputs = layer_module( 2025-09-07T08:03:49.9694637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9694983Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9695352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9695716Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9696073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9696427Z return func(*args, **kwargs) 2025-09-07T08:03:49.9696784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9697144Z self_outputs = self.self( 2025-09-07T08:03:49.9697491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9697834Z return func(*args, **kwargs) 2025-09-07T08:03:49.9698189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9698596Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9698766Z 2025-09-07T08:03:49.9698853Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9699051Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9700483Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9700698Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9700896Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9721128Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9721489Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9721699Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9721891Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9722091Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9722287Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9722484Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9722677Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9722905Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9723270Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9723594Z return mod(**inputs) 2025-09-07T08:03:49.9723978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9724339Z outputs = self.bert( 2025-09-07T08:03:49.9724686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9725058Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9725430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9725795Z layer_outputs = layer_module( 2025-09-07T08:03:49.9726124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9726475Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9726851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9727238Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9727612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9727975Z return func(*args, **kwargs) 2025-09-07T08:03:49.9728332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9728703Z self_outputs = self.self( 2025-09-07T08:03:49.9729055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9729403Z return func(*args, **kwargs) 2025-09-07T08:03:49.9729770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9730191Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9730370Z 2025-09-07T08:03:49.9730464Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9730670Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9730870Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9731072Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9731271Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9731468Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9731667Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9731871Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9732068Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9732267Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9732459Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9732659Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9732859Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9733089Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9733485Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9733804Z return mod(**inputs) 2025-09-07T08:03:49.9734158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9734526Z outputs = self.bert( 2025-09-07T08:03:49.9734895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9735267Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9735633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9736004Z layer_outputs = layer_module( 2025-09-07T08:03:49.9736346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9736690Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9737065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9737447Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9737818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9738173Z return func(*args, **kwargs) 2025-09-07T08:03:49.9738526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9738893Z self_outputs = self.self( 2025-09-07T08:03:49.9739244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9739598Z return func(*args, **kwargs) 2025-09-07T08:03:49.9739949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9740374Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9740563Z 2025-09-07T08:03:49.9740642Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9740851Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9741043Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9741242Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9741448Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9741641Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9741831Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9742029Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9742227Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9742425Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9742622Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9742814Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9743011Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9743246Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9743595Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9743897Z return mod(**inputs) 2025-09-07T08:03:49.9744242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9744606Z outputs = self.bert( 2025-09-07T08:03:49.9744952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9745313Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9745677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9746041Z layer_outputs = layer_module( 2025-09-07T08:03:49.9746379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9746735Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9747146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9747528Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9747901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9748294Z return func(*args, **kwargs) 2025-09-07T08:03:49.9748641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9749007Z self_outputs = self.self( 2025-09-07T08:03:49.9749361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9749720Z return func(*args, **kwargs) 2025-09-07T08:03:49.9750075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9750489Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9750675Z 2025-09-07T08:03:49.9750753Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9750961Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9751163Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9751364Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9751676Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9751872Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9752074Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9752268Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9752455Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9752653Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9752851Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9753047Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9753237Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9753469Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9753820Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9754135Z return mod(**inputs) 2025-09-07T08:03:49.9754473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9754840Z outputs = self.bert( 2025-09-07T08:03:49.9755185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9755553Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9755914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9756272Z layer_outputs = layer_module( 2025-09-07T08:03:49.9756604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9756953Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9757321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9757686Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9758057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9758411Z return func(*args, **kwargs) 2025-09-07T08:03:49.9758765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9759128Z self_outputs = self.self( 2025-09-07T08:03:49.9759463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9759814Z return func(*args, **kwargs) 2025-09-07T08:03:49.9760196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9760610Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9760783Z 2025-09-07T08:03:49.9760867Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9761062Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9761299Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9761499Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9761700Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9761887Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9762084Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9762282Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9762483Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9762673Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9762871Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9763069Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9763270Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9763495Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9763844Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9764159Z return mod(**inputs) 2025-09-07T08:03:49.9764507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9764943Z outputs = self.bert( 2025-09-07T08:03:49.9765278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9765639Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9765990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9766344Z layer_outputs = layer_module( 2025-09-07T08:03:49.9766664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9767005Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9767361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9767729Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9768095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9768507Z return func(*args, **kwargs) 2025-09-07T08:03:49.9768860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9769221Z self_outputs = self.self( 2025-09-07T08:03:49.9769574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9769929Z return func(*args, **kwargs) 2025-09-07T08:03:49.9770285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9770709Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9770886Z 2025-09-07T08:03:49.9770975Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9771184Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9771378Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9771577Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9771777Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9771975Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9772167Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9772367Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9772567Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9772765Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9772954Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9773231Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9773434Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9773663Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9774002Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9774371Z return mod(**inputs) 2025-09-07T08:03:49.9774729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9775100Z outputs = self.bert( 2025-09-07T08:03:49.9775445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9775829Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9776197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9776573Z layer_outputs = layer_module( 2025-09-07T08:03:49.9776918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9777267Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9777643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9778023Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9778402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9778770Z return func(*args, **kwargs) 2025-09-07T08:03:49.9779124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9779495Z self_outputs = self.self( 2025-09-07T08:03:49.9779845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9780210Z return func(*args, **kwargs) 2025-09-07T08:03:49.9780564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9780987Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9781178Z 2025-09-07T08:03:49.9781255Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9781465Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9781671Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9781867Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9782071Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9782275Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9782478Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9782674Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9782874Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9783076Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9783282Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9783478Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9783680Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9783911Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9784269Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9784581Z return mod(**inputs) 2025-09-07T08:03:49.9784934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9785300Z outputs = self.bert( 2025-09-07T08:03:49.9785647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9786016Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9786407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9786779Z layer_outputs = layer_module( 2025-09-07T08:03:49.9787110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9787461Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9787852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9788224Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9788585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9788946Z return func(*args, **kwargs) 2025-09-07T08:03:49.9789300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9789659Z self_outputs = self.self( 2025-09-07T08:03:49.9790007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9790356Z return func(*args, **kwargs) 2025-09-07T08:03:49.9790705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9791127Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9791300Z 2025-09-07T08:03:49.9791385Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9791589Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9791782Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9791982Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9792178Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9792372Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9792561Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9792758Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9792954Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9793157Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9793345Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9793540Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9793736Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9793961Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9794303Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9794613Z return mod(**inputs) 2025-09-07T08:03:49.9794960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9795322Z outputs = self.bert( 2025-09-07T08:03:49.9795658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9796027Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9796391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9796758Z layer_outputs = layer_module( 2025-09-07T08:03:49.9797094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9797433Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9797807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9798177Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9798543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9798888Z return func(*args, **kwargs) 2025-09-07T08:03:49.9799242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9799603Z self_outputs = self.self( 2025-09-07T08:03:49.9799985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9800413Z return func(*args, **kwargs) 2025-09-07T08:03:49.9800761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9801215Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9801397Z 2025-09-07T08:03:49.9801473Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9801677Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9801877Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9802068Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9802269Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9802470Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9802673Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9802862Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9803061Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9803257Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9803456Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9803647Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9803850Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9804079Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9804422Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9804726Z return mod(**inputs) 2025-09-07T08:03:49.9805060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9805407Z outputs = self.bert( 2025-09-07T08:03:49.9805741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9806091Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9806444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9806800Z layer_outputs = layer_module( 2025-09-07T08:03:49.9807126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9807468Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9807818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9808182Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9808535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9808881Z return func(*args, **kwargs) 2025-09-07T08:03:49.9809226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9809572Z self_outputs = self.self( 2025-09-07T08:03:49.9809910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9810253Z return func(*args, **kwargs) 2025-09-07T08:03:49.9810600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9810999Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9811179Z 2025-09-07T08:03:49.9811253Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9811457Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9811656Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9811845Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9812044Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9812235Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9812465Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9812660Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9812846Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9813041Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9813235Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9814106Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9814296Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9814524Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9814867Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9815176Z return mod(**inputs) 2025-09-07T08:03:49.9815513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9815873Z outputs = self.bert( 2025-09-07T08:03:49.9816216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9816576Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9816924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9817278Z layer_outputs = layer_module( 2025-09-07T08:03:49.9817609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9817950Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9818313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9818670Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9819031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9819376Z return func(*args, **kwargs) 2025-09-07T08:03:49.9819718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9820072Z self_outputs = self.self( 2025-09-07T08:03:49.9820405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9820754Z return func(*args, **kwargs) 2025-09-07T08:03:49.9821098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9821510Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9821679Z 2025-09-07T08:03:49.9821753Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9821953Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9822150Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9822346Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9822541Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9822730Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9822923Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9823118Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9823313Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9823500Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9823697Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9823895Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9824089Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9824305Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9824647Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9824952Z return mod(**inputs) 2025-09-07T08:03:49.9825293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9825639Z outputs = self.bert( 2025-09-07T08:03:49.9826010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9826373Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9826730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9827114Z layer_outputs = layer_module( 2025-09-07T08:03:49.9827436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9827777Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9828139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9828507Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9828856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9829205Z return func(*args, **kwargs) 2025-09-07T08:03:49.9829552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9829904Z self_outputs = self.self( 2025-09-07T08:03:49.9830244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9830585Z return func(*args, **kwargs) 2025-09-07T08:03:49.9830931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9831337Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9831507Z 2025-09-07T08:03:49.9831590Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9831781Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9831976Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9832166Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9832361Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9832554Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9832738Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9832930Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9833125Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9833318Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9833503Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9833695Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9833888Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9834106Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9834438Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9834741Z return mod(**inputs) 2025-09-07T08:03:49.9835081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9835435Z outputs = self.bert( 2025-09-07T08:03:49.9835762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9836123Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9836477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9836838Z layer_outputs = layer_module( 2025-09-07T08:03:49.9837164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9837496Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9837859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9838224Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9838615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9838954Z return func(*args, **kwargs) 2025-09-07T08:03:49.9839298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9839648Z self_outputs = self.self( 2025-09-07T08:03:49.9840016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9840369Z return func(*args, **kwargs) 2025-09-07T08:03:49.9840707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9841119Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9841303Z 2025-09-07T08:03:49.9841382Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9841588Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9841793Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9841991Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9842192Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9842396Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9842597Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9842790Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9842992Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9843193Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9843393Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9843583Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9843785Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9844016Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9844358Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9844658Z return mod(**inputs) 2025-09-07T08:03:49.9845004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9845362Z outputs = self.bert( 2025-09-07T08:03:49.9845701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9846062Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9846424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9846786Z layer_outputs = layer_module( 2025-09-07T08:03:49.9847120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9847463Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9847823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9848194Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9848557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9848905Z return func(*args, **kwargs) 2025-09-07T08:03:49.9849257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9849608Z self_outputs = self.self( 2025-09-07T08:03:49.9849945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9850289Z return func(*args, **kwargs) 2025-09-07T08:03:49.9850633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9851032Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9851212Z 2025-09-07T08:03:49.9851291Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9851496Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9851728Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9851923Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9852108Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9852300Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9852494Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9852711Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9852895Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9853091Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9853284Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9853472Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9853656Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9853877Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9854216Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9854514Z return mod(**inputs) 2025-09-07T08:03:49.9854846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9855197Z outputs = self.bert( 2025-09-07T08:03:49.9855532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9855892Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9856231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9856587Z layer_outputs = layer_module( 2025-09-07T08:03:49.9856914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9857252Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9857607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9857964Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9858319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9858667Z return func(*args, **kwargs) 2025-09-07T08:03:49.9859008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9859358Z self_outputs = self.self( 2025-09-07T08:03:49.9859684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9860024Z return func(*args, **kwargs) 2025-09-07T08:03:49.9860365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9860765Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9860931Z 2025-09-07T08:03:49.9861010Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9861200Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9861391Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9861586Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9861778Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9861960Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9862152Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9862341Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9862533Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9862721Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9862909Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9863098Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9863289Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9863503Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9863842Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9864173Z return mod(**inputs) 2025-09-07T08:03:49.9864509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9864857Z outputs = self.bert( 2025-09-07T08:03:49.9865203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9865606Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9865966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9866329Z layer_outputs = layer_module( 2025-09-07T08:03:49.9866656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9866999Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9867367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9867741Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9868095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9868512Z return func(*args, **kwargs) 2025-09-07T08:03:49.9868870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9869229Z self_outputs = self.self( 2025-09-07T08:03:49.9869573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9869916Z return func(*args, **kwargs) 2025-09-07T08:03:49.9870269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9870680Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9870852Z 2025-09-07T08:03:49.9870941Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9871148Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9871338Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9871537Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9871736Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9871925Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9872113Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9872304Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9872498Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9872689Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9872875Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9873071Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9873265Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9873487Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9873821Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9874124Z return mod(**inputs) 2025-09-07T08:03:49.9874462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9874818Z outputs = self.bert( 2025-09-07T08:03:49.9875149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9875513Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9875873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9876230Z layer_outputs = layer_module( 2025-09-07T08:03:49.9876555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9876891Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9877317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9877686Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9878045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9878436Z return func(*args, **kwargs) 2025-09-07T08:03:49.9878784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9879140Z self_outputs = self.self( 2025-09-07T08:03:49.9879482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9879828Z return func(*args, **kwargs) 2025-09-07T08:03:49.9880167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9880579Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9880754Z 2025-09-07T08:03:49.9880828Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9881029Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9881223Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9881415Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9881609Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9881805Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9881995Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9882182Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9882373Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9882566Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9882758Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9882942Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9883140Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9883365Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9883699Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9883993Z return mod(**inputs) 2025-09-07T08:03:49.9884331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9884687Z outputs = self.bert( 2025-09-07T08:03:49.9885017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9885366Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9885719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9886069Z layer_outputs = layer_module( 2025-09-07T08:03:49.9886393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9886736Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9887087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9887448Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9887810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9888157Z return func(*args, **kwargs) 2025-09-07T08:03:49.9888498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9888846Z self_outputs = self.self( 2025-09-07T08:03:49.9889179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9889526Z return func(*args, **kwargs) 2025-09-07T08:03:49.9889899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9890301Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9890479Z 2025-09-07T08:03:49.9890556Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9890760Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9890993Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9891186Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9891374Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9891570Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9891765Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9891961Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9892148Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9892345Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9892541Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9892731Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9892921Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9893143Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9893480Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9893785Z return mod(**inputs) 2025-09-07T08:03:49.9894120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9894472Z outputs = self.bert( 2025-09-07T08:03:49.9894807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9895167Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9895521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9895868Z layer_outputs = layer_module( 2025-09-07T08:03:49.9896198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9896541Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9896900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9897256Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9897614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9897962Z return func(*args, **kwargs) 2025-09-07T08:03:49.9898309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9898658Z self_outputs = self.self( 2025-09-07T08:03:49.9898990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9899335Z return func(*args, **kwargs) 2025-09-07T08:03:49.9899676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9900080Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9900247Z 2025-09-07T08:03:49.9900328Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9900524Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9900719Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9900914Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9901111Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9901295Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9901483Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9901673Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9901867Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9902052Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9902242Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9902466Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9902662Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9902877Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9903212Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9903566Z return mod(**inputs) 2025-09-07T08:03:49.9903905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9904253Z outputs = self.bert( 2025-09-07T08:03:49.9904589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9904950Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9905308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9905661Z layer_outputs = layer_module( 2025-09-07T08:03:49.9905984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9906326Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9906687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9907051Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9907402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9907747Z return func(*args, **kwargs) 2025-09-07T08:03:49.9908097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9908450Z self_outputs = self.self( 2025-09-07T08:03:49.9908790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9909132Z return func(*args, **kwargs) 2025-09-07T08:03:49.9909481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9909887Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9910061Z 2025-09-07T08:03:49.9910144Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9910342Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9910528Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9910722Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9910915Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9911110Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9911298Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9911487Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9911681Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9911873Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9912065Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9912259Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9912453Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9912672Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9913001Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9913303Z return mod(**inputs) 2025-09-07T08:03:49.9913646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9913997Z outputs = self.bert( 2025-09-07T08:03:49.9914326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9914688Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9915041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9915437Z layer_outputs = layer_module( 2025-09-07T08:03:49.9915768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9916104Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9916468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9916867Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9917225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9917574Z return func(*args, **kwargs) 2025-09-07T08:03:49.9917912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9918261Z self_outputs = self.self( 2025-09-07T08:03:49.9918606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9918962Z return func(*args, **kwargs) 2025-09-07T08:03:49.9919302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9919715Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9919898Z 2025-09-07T08:03:49.9919973Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9920175Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9920372Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9920562Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9920754Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9920948Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9921145Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9921329Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9921525Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9921720Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9921914Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9922103Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9922293Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9922515Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:49.9922857Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:03:49.9923151Z return mod(**inputs) 2025-09-07T08:03:49.9923491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:03:49.9923846Z outputs = self.bert( 2025-09-07T08:03:49.9924186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:03:49.9924538Z encoder_outputs = self.encoder( 2025-09-07T08:03:49.9924888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:03:49.9925242Z layer_outputs = layer_module( 2025-09-07T08:03:49.9925566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:03:49.9925902Z return super().__call__(*args, **kwargs) 2025-09-07T08:03:49.9926255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:03:49.9926615Z self_attention_outputs = self.attention( 2025-09-07T08:03:49.9926970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9927319Z return func(*args, **kwargs) 2025-09-07T08:03:49.9927666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:03:49.9928011Z self_outputs = self.self( 2025-09-07T08:03:49.9928376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:03:49.9928725Z return func(*args, **kwargs) 2025-09-07T08:03:49.9929070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:03:49.9929492Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:03:49.9929669Z 2025-09-07T08:03:49.9929742Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9929937Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9930138Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9930329Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9930512Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9930704Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9930895Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9931090Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9931282Z cudagraph partition due to non gpu ops 2025-09-07T08:03:49.9931480Z cudagraph partition due to non gpu ops 2025-09-07T08:04:03.0793006Z pass 2025-09-07T08:04:03.0793499Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:05.0722143Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:04:05.0723538Z import pynvml # type: ignore[import] 2025-09-07T08:04:06.7876861Z 2025-09-07T08:04:07.9318079Z loading model: 0it [00:00, ?it/s]BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:04:07.9321795Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:04:07.9322555Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:04:07.9323145Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:04:07.9324300Z WARNING:transformers.modeling_utils:BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:04:07.9325395Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:04:07.9326126Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:04:07.9326711Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:04:09.2912326Z Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:04:09.2913323Z WARNING:transformers.models.big_bird.modeling_big_bird:Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:04:10.3425335Z 2025-09-07T08:04:10.3425884Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:04:10.3429864Z cpu eval hf_BigBird 2025-09-07T08:04:11.7173267Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:13.2270273Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:14.7335406Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:25.0285590Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:25.0286111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1897, in forward 2025-09-07T08:04:25.0286605Z attention_mask = torch.ones(((batch_size, seq_length + past_key_values_length)), device=device) 2025-09-07T08:04:25.0286842Z 2025-09-07T08:04:26.1121050Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:26.1121702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2090, in torch_dynamo_resume_in__pad_to_block_size_at_2085 2025-09-07T08:04:26.1122249Z input_ids = nn.functional.pad(input_ids, (0, padding_len), value=pad_token_id) 2025-09-07T08:04:26.1122643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:04:26.1123010Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:04:26.1123191Z 2025-09-07T08:04:26.1123298Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:26.1123780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2103, in torch_dynamo_resume_in__pad_to_block_size_at_2085 2025-09-07T08:04:26.1124257Z attention_mask = nn.functional.pad( 2025-09-07T08:04:26.1124567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:04:26.1124901Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:04:26.1125050Z 2025-09-07T08:04:26.1125149Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:26.1125618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2106, in torch_dynamo_resume_in__pad_to_block_size_at_2085 2025-09-07T08:04:26.1126183Z token_type_ids = nn.functional.pad(token_type_ids, (0, padding_len), value=0) # pad with token_type_id = 0 2025-09-07T08:04:26.1126595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:04:26.1126918Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:04:26.1127065Z 2025-09-07T08:04:28.4325725Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:28.4326387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1983, in torch_dynamo_resume_in_forward_at_1935 2025-09-07T08:04:28.4326961Z embedding_output = self.embeddings( 2025-09-07T08:04:28.4327418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 292, in forward 2025-09-07T08:04:28.4327876Z embeddings += position_embeddings 2025-09-07T08:04:28.4328028Z 2025-09-07T08:04:28.4328123Z cudagraph partition due to non gpu ops 2025-09-07T08:04:28.4328348Z cudagraph partition due to non gpu ops 2025-09-07T08:04:28.4328561Z cudagraph partition due to non gpu ops 2025-09-07T08:04:28.4328803Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:28.4329311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1947, in torch_dynamo_resume_in_forward_at_1935 2025-09-07T08:04:28.4329921Z blocked_encoder_mask, band_mask, from_mask, to_mask = self.create_masks_for_block_sparse_attn( 2025-09-07T08:04:28.4330823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2060, in create_masks_for_block_sparse_attn 2025-09-07T08:04:28.4331352Z band_mask = create_band_mask_from_inputs(blocked_encoder_mask, blocked_encoder_mask) 2025-09-07T08:04:28.4331854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2052, in create_band_mask_from_inputs 2025-09-07T08:04:28.4332383Z exp_blocked_to_pad = torch.cat( 2025-09-07T08:04:28.4332505Z 2025-09-07T08:04:28.4332615Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:28.4333066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1947, in torch_dynamo_resume_in_forward_at_1935 2025-09-07T08:04:28.4333595Z blocked_encoder_mask, band_mask, from_mask, to_mask = self.create_masks_for_block_sparse_attn( 2025-09-07T08:04:28.4334117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2060, in create_masks_for_block_sparse_attn 2025-09-07T08:04:28.4334623Z band_mask = create_band_mask_from_inputs(blocked_encoder_mask, blocked_encoder_mask) 2025-09-07T08:04:28.4335122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2055, in create_band_mask_from_inputs 2025-09-07T08:04:28.4335631Z band_mask = torch.einsum("blq,blk->blqk", from_blocked_mask[:, 2:-2], exp_blocked_to_pad) 2025-09-07T08:04:28.4335837Z 2025-09-07T08:04:33.2175009Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:33.2175595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:04:33.2175984Z self.query(hidden_states) 2025-09-07T08:04:33.2176112Z 2025-09-07T08:04:33.2176194Z cudagraph partition due to non gpu ops 2025-09-07T08:04:33.2176406Z cudagraph partition due to non gpu ops 2025-09-07T08:04:33.2176636Z cudagraph partition due to non gpu ops 2025-09-07T08:04:33.2176829Z cudagraph partition due to non gpu ops 2025-09-07T08:04:33.2177026Z cudagraph partition due to non gpu ops 2025-09-07T08:04:33.2177220Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2714370Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2714711Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2714918Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2715113Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2715314Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2715508Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2715703Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2715894Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2716087Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2716284Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2716484Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2716686Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2716896Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2717091Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2717286Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2717512Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2718132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 647, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2718670Z to_mask.new_ones([bsz, 1, 1, n_rand_blocks * to_block_size]), 2025-09-07T08:04:37.2718838Z 2025-09-07T08:04:37.2718913Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2719125Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2719392Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2720148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 602, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2720750Z first_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, 0], key_layer, ndim=4) 2025-09-07T08:04:37.2721280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 494, in torch_bmm_nd_transpose 2025-09-07T08:04:37.2721877Z inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:]).transpose(1, 2) 2025-09-07T08:04:37.2722097Z 2025-09-07T08:04:37.2722183Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2722389Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2722613Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2723125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 642, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2723720Z second_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, 1], second_key_mat, ndim=4) 2025-09-07T08:04:37.2724250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 494, in torch_bmm_nd_transpose 2025-09-07T08:04:37.2724742Z inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:]).transpose(1, 2) 2025-09-07T08:04:37.2724947Z 2025-09-07T08:04:37.2725023Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2725224Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2725450Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2725956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 698, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2726429Z first_band_product = torch.einsum( 2025-09-07T08:04:37.2726552Z 2025-09-07T08:04:37.2726652Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2727153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 704, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2727634Z last_band_product = torch.einsum( 2025-09-07T08:04:37.2727757Z 2025-09-07T08:04:37.2727838Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2728033Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2728246Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2728744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 776, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2729361Z second_last_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, -2], second_last_key_mat, ndim=4) 2025-09-07T08:04:37.2729907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 494, in torch_bmm_nd_transpose 2025-09-07T08:04:37.2730399Z inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:]).transpose(1, 2) 2025-09-07T08:04:37.2730599Z 2025-09-07T08:04:37.2730674Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2730873Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2731102Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2731601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 807, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2732169Z last_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, -1], key_layer, ndim=4) 2025-09-07T08:04:37.2732672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 494, in torch_bmm_nd_transpose 2025-09-07T08:04:37.2733197Z inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:]).transpose(1, 2) 2025-09-07T08:04:37.2733409Z 2025-09-07T08:04:37.2733510Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2734012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 653, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2734590Z rand_mask.new_ones([bsz, n_heads, from_block_size, 4 * to_block_size]), 2025-09-07T08:04:37.2734773Z 2025-09-07T08:04:37.2734850Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2735054Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2735254Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2735450Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2735641Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2735841Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2736038Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2736265Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2736777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 602, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2737340Z first_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, 0], key_layer, ndim=4) 2025-09-07T08:04:37.2737856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 494, in torch_bmm_nd_transpose 2025-09-07T08:04:37.2738348Z inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:]).transpose(1, 2) 2025-09-07T08:04:37.2738548Z 2025-09-07T08:04:37.2738655Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2739157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 611, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2739704Z first_context_layer = self.torch_bmm_nd(first_attn_weights, value_layer, ndim=4) 2025-09-07T08:04:37.2740172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:04:37.2740665Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:04:37.2740899Z 2025-09-07T08:04:37.2741000Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2741504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 574, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2741986Z rand_attn = np.stack(rand_attn, axis=0) 2025-09-07T08:04:37.2742117Z 2025-09-07T08:04:37.2742217Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2742723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 643, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2743199Z second_seq_pad = torch.cat( 2025-09-07T08:04:37.2743315Z 2025-09-07T08:04:37.2743422Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2743932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 777, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2744402Z second_last_seq_pad = torch.cat( 2025-09-07T08:04:37.2744530Z 2025-09-07T08:04:37.2744627Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2745127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 577, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2745651Z rand_attn = torch.cat([rand_attn for _ in range(batch_size)], dim=0) 2025-09-07T08:04:37.2745827Z 2025-09-07T08:04:37.2745966Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2746457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 579, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2746986Z rand_mask = self._create_rand_mask_from_inputs( 2025-09-07T08:04:37.2747428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in _create_rand_mask_from_inputs 2025-09-07T08:04:37.2747933Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:04:37.2748401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in 2025-09-07T08:04:37.2748858Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:04:37.2749067Z 2025-09-07T08:04:37.2749169Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2749672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 579, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2750172Z rand_mask = self._create_rand_mask_from_inputs( 2025-09-07T08:04:37.2750614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in _create_rand_mask_from_inputs 2025-09-07T08:04:37.2751110Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:04:37.2751564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in 2025-09-07T08:04:37.2752024Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:04:37.2752232Z 2025-09-07T08:04:37.2752334Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2752837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 579, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2753333Z rand_mask = self._create_rand_mask_from_inputs( 2025-09-07T08:04:37.2753766Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in _create_rand_mask_from_inputs 2025-09-07T08:04:37.2754262Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:04:37.2754722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in 2025-09-07T08:04:37.2755183Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:04:37.2755382Z 2025-09-07T08:04:37.2755495Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2755988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 579, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2756477Z rand_mask = self._create_rand_mask_from_inputs( 2025-09-07T08:04:37.2756915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in _create_rand_mask_from_inputs 2025-09-07T08:04:37.2757413Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:04:37.2757871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in 2025-09-07T08:04:37.2758328Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:04:37.2758526Z 2025-09-07T08:04:37.2758626Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2759179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 579, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2759676Z rand_mask = self._create_rand_mask_from_inputs( 2025-09-07T08:04:37.2760897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in _create_rand_mask_from_inputs 2025-09-07T08:04:37.2761392Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:04:37.2761592Z 2025-09-07T08:04:37.2761671Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2761882Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2762117Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2762623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 620, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2763109Z second_key_mat = torch.cat( 2025-09-07T08:04:37.2763226Z 2025-09-07T08:04:37.2763324Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2763828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 642, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2764420Z second_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, 1], second_key_mat, ndim=4) 2025-09-07T08:04:37.2764945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:04:37.2765353Z return torch.bmm( 2025-09-07T08:04:37.2765449Z 2025-09-07T08:04:37.2765546Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2766055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 630, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2766538Z second_value_mat = torch.cat( 2025-09-07T08:04:37.2766657Z 2025-09-07T08:04:37.2766762Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2767263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 754, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2767733Z second_last_key_mat = torch.cat( 2025-09-07T08:04:37.2767863Z 2025-09-07T08:04:37.2767960Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2768556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 776, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2769168Z second_last_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, -2], second_last_key_mat, ndim=4) 2025-09-07T08:04:37.2769715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:04:37.2770112Z return torch.bmm( 2025-09-07T08:04:37.2770216Z 2025-09-07T08:04:37.2770315Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2770822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 764, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2771306Z second_last_value_mat = torch.cat( 2025-09-07T08:04:37.2771431Z 2025-09-07T08:04:37.2771537Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2772031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 602, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2772680Z first_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, 0], key_layer, ndim=4) 2025-09-07T08:04:37.2773193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:04:37.2773599Z return torch.bmm( 2025-09-07T08:04:37.2773695Z 2025-09-07T08:04:37.2773886Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2774389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 807, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2774957Z last_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, -1], key_layer, ndim=4) 2025-09-07T08:04:37.2775459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:04:37.2775859Z return torch.bmm( 2025-09-07T08:04:37.2775960Z 2025-09-07T08:04:37.2776058Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2776560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 651, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2777034Z second_rand_pad = torch.cat( 2025-09-07T08:04:37.2777157Z 2025-09-07T08:04:37.2777256Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2777760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 785, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2778237Z second_last_rand_pad = torch.cat( 2025-09-07T08:04:37.2778361Z 2025-09-07T08:04:37.2778466Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2778963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 665, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2779522Z second_context_layer = self.torch_bmm_nd(second_attn_weights, second_value_mat, ndim=4) 2025-09-07T08:04:37.2779998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:04:37.2780496Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:04:37.2780727Z 2025-09-07T08:04:37.2780827Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2781324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 799, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2781914Z second_last_context_layer = self.torch_bmm_nd(second_last_attn_weights, second_last_value_mat, ndim=4) 2025-09-07T08:04:37.2782415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:04:37.2782909Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:04:37.2783140Z 2025-09-07T08:04:37.2783239Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2783753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 676, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2784239Z exp_blocked_key_matrix = torch.cat( 2025-09-07T08:04:37.2784369Z 2025-09-07T08:04:37.2784466Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2784977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 679, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2785474Z exp_blocked_value_matrix = torch.cat( 2025-09-07T08:04:37.2785612Z 2025-09-07T08:04:37.2785716Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2785930Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2786159Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2786669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 611, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2787239Z first_context_layer = self.torch_bmm_nd(first_attn_weights, value_layer, ndim=4) 2025-09-07T08:04:37.2787708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:04:37.2788194Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:04:37.2788414Z 2025-09-07T08:04:37.2788498Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2788719Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2789221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 813, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2789769Z last_context_layer = self.torch_bmm_nd(last_attn_weights, value_layer, ndim=4) 2025-09-07T08:04:37.2790236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:04:37.2790727Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:04:37.2790947Z 2025-09-07T08:04:37.2791025Z cudagraph partition due to non gpu ops 2025-09-07T08:04:37.2791255Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2791758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 687, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2792352Z inner_band_product = self.torch_bmm_nd_transpose(middle_query_matrix, exp_blocked_key_matrix, ndim=5) 2025-09-07T08:04:37.2792879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:04:37.2793287Z return torch.bmm( 2025-09-07T08:04:37.2793392Z 2025-09-07T08:04:37.2793493Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2794001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 710, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2794520Z inner_band_product += (1.0 - band_mask) * attn_mask_penalty 2025-09-07T08:04:37.2794686Z 2025-09-07T08:04:37.2794791Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2795286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 698, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2795769Z first_band_product = torch.einsum( 2025-09-07T08:04:37.2795902Z 2025-09-07T08:04:37.2796012Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2796507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 704, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2796989Z last_band_product = torch.einsum( 2025-09-07T08:04:37.2797114Z 2025-09-07T08:04:37.2797212Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2797719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 711, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2798282Z first_band_product += (1.0 - to_mask[:, :, :, :to_block_size].unsqueeze(3)) * attn_mask_penalty 2025-09-07T08:04:37.2798520Z 2025-09-07T08:04:37.2798620Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2799120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 712, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2799702Z last_band_product += (1.0 - to_mask[:, :, :, -to_block_size:].unsqueeze(3)) * attn_mask_penalty 2025-09-07T08:04:37.2799902Z 2025-09-07T08:04:37.2799999Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2800500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 693, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2801090Z rand_band_product = self.torch_bmm_nd_transpose(middle_query_matrix, gathered_key[:, :, 1:-1], ndim=5) 2025-09-07T08:04:37.2801624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:04:37.2802033Z return torch.bmm( 2025-09-07T08:04:37.2802128Z 2025-09-07T08:04:37.2802226Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2802723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 713, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2803251Z rand_band_product += (1.0 - rand_mask[:, :, 1:-1]) * attn_mask_penalty 2025-09-07T08:04:37.2803426Z 2025-09-07T08:04:37.2803530Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2804031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 716, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2804501Z band_product = torch.cat( 2025-09-07T08:04:37.2804623Z 2025-09-07T08:04:37.2804723Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2805219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 727, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2805693Z context_layer = self.torch_bmm_nd( 2025-09-07T08:04:37.2806088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:04:37.2806574Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:04:37.2806800Z 2025-09-07T08:04:37.2806898Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2807396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 734, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2807874Z context_layer += self.torch_bmm_nd( 2025-09-07T08:04:37.2808273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:04:37.2808759Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:04:37.2808980Z 2025-09-07T08:04:37.2809078Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2809576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 740, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2810055Z context_layer += torch.einsum( 2025-09-07T08:04:37.2810177Z 2025-09-07T08:04:37.2810282Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2810781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 743, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2811280Z context_layer += torch.einsum( 2025-09-07T08:04:37.2811406Z 2025-09-07T08:04:37.2811507Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2812012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 743, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2812516Z context_layer += torch.einsum( 2025-09-07T08:04:37.2812635Z 2025-09-07T08:04:37.2812739Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2813228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 817, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2813744Z context_layer = torch.cat( 2025-09-07T08:04:37.2813867Z 2025-09-07T08:04:37.2813963Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.2814469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 821, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:04:37.2815018Z context_layer = context_layer.view((bsz, n_heads, from_seq_len, -1)) * from_mask 2025-09-07T08:04:37.2815221Z 2025-09-07T08:04:43.7600368Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:43.7601045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 478, in torch_dynamo_resume_in_forward_at_455 2025-09-07T08:04:43.7601686Z context_layer = context_layer.contiguous().view(batch_size, from_seq_length, -1) 2025-09-07T08:04:43.7601944Z 2025-09-07T08:04:43.9841961Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:43.9842746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:04:43.9843399Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:04:43.9843943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:04:43.9844427Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:43.9844615Z 2025-09-07T08:04:43.9844724Z cudagraph partition due to non gpu ops 2025-09-07T08:04:43.9844966Z cudagraph partition due to non gpu ops 2025-09-07T08:04:44.2693861Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:44.2694444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:44.2694927Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:44.2695347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:44.2695722Z return forward_fn(*input_tensors) 2025-09-07T08:04:44.2696150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:44.2696600Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:44.2697016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:04:44.2697415Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:44.2697549Z 2025-09-07T08:04:44.2697631Z cudagraph partition due to non gpu ops 2025-09-07T08:04:44.2697838Z cudagraph partition due to non gpu ops 2025-09-07T08:04:44.2698066Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:44.2698524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:44.2698968Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:44.2699611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:44.2700014Z return forward_fn(*input_tensors) 2025-09-07T08:04:44.2700416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:44.2700987Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:44.2701398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:04:44.2701809Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:44.2702177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:04:44.2702611Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:04:44.2702832Z 2025-09-07T08:04:44.2702918Z cudagraph partition due to non gpu ops 2025-09-07T08:04:44.2703113Z cudagraph partition due to non gpu ops 2025-09-07T08:04:44.5561141Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:44.5563208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:04:44.5563686Z self.query(hidden_states) 2025-09-07T08:04:44.5563821Z 2025-09-07T08:04:44.5563909Z cudagraph partition due to non gpu ops 2025-09-07T08:04:44.5564130Z cudagraph partition due to non gpu ops 2025-09-07T08:04:44.5564328Z cudagraph partition due to non gpu ops 2025-09-07T08:04:44.5564526Z cudagraph partition due to non gpu ops 2025-09-07T08:04:44.5564721Z cudagraph partition due to non gpu ops 2025-09-07T08:04:44.5564916Z cudagraph partition due to non gpu ops 2025-09-07T08:04:45.5421454Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:45.5422075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:04:45.5422587Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:04:45.5423021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:04:45.5423431Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:45.5423571Z 2025-09-07T08:04:45.5423655Z cudagraph partition due to non gpu ops 2025-09-07T08:04:45.5423863Z cudagraph partition due to non gpu ops 2025-09-07T08:04:45.7798194Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:45.7798767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:45.7799237Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:45.7799666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:45.7800054Z return forward_fn(*input_tensors) 2025-09-07T08:04:45.7800456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:45.7800922Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:45.7801345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:04:45.7801728Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:45.7801872Z 2025-09-07T08:04:45.7801954Z cudagraph partition due to non gpu ops 2025-09-07T08:04:45.7802161Z cudagraph partition due to non gpu ops 2025-09-07T08:04:45.7802389Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:45.7803159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:45.7803616Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:45.7804003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:45.7804484Z return forward_fn(*input_tensors) 2025-09-07T08:04:45.7804896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:45.7805337Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:45.7805741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:04:45.7806148Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:45.7806520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:04:45.7806952Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:04:45.7807170Z 2025-09-07T08:04:45.7807254Z cudagraph partition due to non gpu ops 2025-09-07T08:04:45.7807449Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.0522956Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:46.0523581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:04:46.0524047Z self.query(hidden_states) 2025-09-07T08:04:46.0524194Z 2025-09-07T08:04:46.0524298Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.0524545Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.0524779Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.0525022Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.0525280Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.0525512Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.2993686Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:46.2994285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:04:46.2994849Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:04:46.2995289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:04:46.2995680Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:46.2995819Z 2025-09-07T08:04:46.2995912Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.2996111Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.5349760Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:46.5350355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:46.5350830Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:46.5351228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:46.5351620Z return forward_fn(*input_tensors) 2025-09-07T08:04:46.5352025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:46.5352471Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:46.5352891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:04:46.5353279Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:46.5353412Z 2025-09-07T08:04:46.5353495Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.5357045Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.5357326Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:46.5357795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:46.5358345Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:46.5358720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:46.5359094Z return forward_fn(*input_tensors) 2025-09-07T08:04:46.5359491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:46.5359943Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:46.5360370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:04:46.5360775Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:46.5361148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:04:46.5361577Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:04:46.5361801Z 2025-09-07T08:04:46.5361885Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.5362087Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.8061240Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:46.8061828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:04:46.8062304Z self.query(hidden_states) 2025-09-07T08:04:46.8062449Z 2025-09-07T08:04:46.8062545Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.8062813Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.8063083Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.8063331Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.8063525Z cudagraph partition due to non gpu ops 2025-09-07T08:04:46.8063709Z cudagraph partition due to non gpu ops 2025-09-07T08:04:47.0541517Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:47.0542052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:04:47.0542545Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:04:47.0542981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:04:47.0543371Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:47.0543508Z 2025-09-07T08:04:47.0543590Z cudagraph partition due to non gpu ops 2025-09-07T08:04:47.0543815Z cudagraph partition due to non gpu ops 2025-09-07T08:04:47.2897665Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:47.2898236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:47.2898725Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:47.2899135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:47.2899521Z return forward_fn(*input_tensors) 2025-09-07T08:04:47.2899926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:47.2900377Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:47.2901070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:04:47.2901468Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:47.2901615Z 2025-09-07T08:04:47.2901697Z cudagraph partition due to non gpu ops 2025-09-07T08:04:47.2901910Z cudagraph partition due to non gpu ops 2025-09-07T08:04:47.2902144Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:47.2902699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:47.2903132Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:47.2903514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:47.2903881Z return forward_fn(*input_tensors) 2025-09-07T08:04:47.2904276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:47.2904718Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:47.2905124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:04:47.2905530Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:47.2905899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:04:47.2906325Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:04:47.2906539Z 2025-09-07T08:04:47.2906621Z cudagraph partition due to non gpu ops 2025-09-07T08:04:47.2906815Z cudagraph partition due to non gpu ops 2025-09-07T08:04:47.5618370Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:47.5618851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:04:47.5619263Z self.query(hidden_states) 2025-09-07T08:04:47.5619381Z 2025-09-07T08:04:47.5619475Z cudagraph partition due to non gpu ops 2025-09-07T08:04:47.5619678Z cudagraph partition due to non gpu ops 2025-09-07T08:04:47.5619884Z cudagraph partition due to non gpu ops 2025-09-07T08:04:47.5620094Z cudagraph partition due to non gpu ops 2025-09-07T08:04:47.5620288Z cudagraph partition due to non gpu ops 2025-09-07T08:04:47.5620474Z cudagraph partition due to non gpu ops 2025-09-07T08:04:47.8083633Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:47.8084329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:04:47.8084874Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:04:47.8085317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:04:47.8085742Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:47.8085897Z 2025-09-07T08:04:47.8085981Z cudagraph partition due to non gpu ops 2025-09-07T08:04:47.8086186Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.0431862Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:48.0432565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:48.0433062Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:48.0433469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:48.0433853Z return forward_fn(*input_tensors) 2025-09-07T08:04:48.0434255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:48.0434933Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:48.0435372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:04:48.0435765Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:48.0436000Z 2025-09-07T08:04:48.0436090Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.0436293Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.0436521Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:48.0436982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:48.0437428Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:48.0437813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:48.0438183Z return forward_fn(*input_tensors) 2025-09-07T08:04:48.0438582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:48.0439023Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:48.0439441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:04:48.0439853Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:48.0440215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:04:48.0440642Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:04:48.0440866Z 2025-09-07T08:04:48.0440942Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.0441145Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.3171126Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:48.3172803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:04:48.3173296Z self.query(hidden_states) 2025-09-07T08:04:48.3173447Z 2025-09-07T08:04:48.3173565Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.3173818Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.3174057Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.3174297Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.3174495Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.3174689Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.5654911Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:48.5655471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:04:48.5655994Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:04:48.5656444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:04:48.5656840Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:48.5657002Z 2025-09-07T08:04:48.5657084Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.5657290Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.8010922Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:48.8011671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:48.8012174Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:48.8012582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:48.8013242Z return forward_fn(*input_tensors) 2025-09-07T08:04:48.8013664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:48.8014099Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:48.8014667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:04:48.8015057Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:48.8015193Z 2025-09-07T08:04:48.8015285Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.8015497Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.8015717Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:48.8016184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:48.8016630Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:48.8017012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:48.8017388Z return forward_fn(*input_tensors) 2025-09-07T08:04:48.8017778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:48.8018218Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:48.8018631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:04:48.8019044Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:48.8019408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:04:48.8019834Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:04:48.8020061Z 2025-09-07T08:04:48.8020136Z cudagraph partition due to non gpu ops 2025-09-07T08:04:48.8020336Z cudagraph partition due to non gpu ops 2025-09-07T08:04:49.0729839Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:49.0730558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:04:49.0731111Z self.query(hidden_states) 2025-09-07T08:04:49.0731235Z 2025-09-07T08:04:49.0731323Z cudagraph partition due to non gpu ops 2025-09-07T08:04:49.0731540Z cudagraph partition due to non gpu ops 2025-09-07T08:04:49.0731741Z cudagraph partition due to non gpu ops 2025-09-07T08:04:49.0731941Z cudagraph partition due to non gpu ops 2025-09-07T08:04:49.0732133Z cudagraph partition due to non gpu ops 2025-09-07T08:04:49.0732336Z cudagraph partition due to non gpu ops 2025-09-07T08:04:49.3187083Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:49.3187822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:04:49.3188438Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:04:49.3188940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:04:49.3189354Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:49.3189503Z 2025-09-07T08:04:49.3189588Z cudagraph partition due to non gpu ops 2025-09-07T08:04:49.3189800Z cudagraph partition due to non gpu ops 2025-09-07T08:04:49.8980805Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:49.8981494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:49.8982056Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:49.8982888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:49.8983315Z return forward_fn(*input_tensors) 2025-09-07T08:04:49.8983722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:49.8986167Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:49.8986591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:04:49.8986985Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:49.8987119Z 2025-09-07T08:04:49.8987207Z cudagraph partition due to non gpu ops 2025-09-07T08:04:49.8987413Z cudagraph partition due to non gpu ops 2025-09-07T08:04:49.8987633Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:49.8988096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:49.8988534Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:49.8988909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:49.8989282Z return forward_fn(*input_tensors) 2025-09-07T08:04:49.8989678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:49.8990114Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:49.8990530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:04:49.8990940Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:49.8991301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:04:49.8991732Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:04:49.8991956Z 2025-09-07T08:04:49.8992033Z cudagraph partition due to non gpu ops 2025-09-07T08:04:49.8992239Z cudagraph partition due to non gpu ops 2025-09-07T08:04:50.1719263Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:50.1721539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:04:50.1721942Z self.query(hidden_states) 2025-09-07T08:04:50.1722062Z 2025-09-07T08:04:50.1722153Z cudagraph partition due to non gpu ops 2025-09-07T08:04:50.1722358Z cudagraph partition due to non gpu ops 2025-09-07T08:04:50.1722561Z cudagraph partition due to non gpu ops 2025-09-07T08:04:50.1722761Z cudagraph partition due to non gpu ops 2025-09-07T08:04:50.1722981Z cudagraph partition due to non gpu ops 2025-09-07T08:04:50.1723170Z cudagraph partition due to non gpu ops 2025-09-07T08:04:50.4167124Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:50.4167731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:04:50.4168366Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:04:50.4168830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:04:50.4169232Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:50.4169392Z 2025-09-07T08:04:50.4169479Z cudagraph partition due to non gpu ops 2025-09-07T08:04:50.4169734Z cudagraph partition due to non gpu ops 2025-09-07T08:04:50.6522902Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:50.6523995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:50.6524524Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:50.6524926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:50.6525405Z return forward_fn(*input_tensors) 2025-09-07T08:04:50.6525809Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:50.6526255Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:50.6526674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:04:50.6527058Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:50.6527193Z 2025-09-07T08:04:50.6527276Z cudagraph partition due to non gpu ops 2025-09-07T08:04:50.6527489Z cudagraph partition due to non gpu ops 2025-09-07T08:04:50.6527719Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:50.6528182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:04:50.6528631Z layer_output = apply_chunking_to_forward( 2025-09-07T08:04:50.6529004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:04:50.6529373Z return forward_fn(*input_tensors) 2025-09-07T08:04:50.6529769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:04:50.6530205Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:50.6530621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:04:50.6531027Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:50.6531398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:04:50.6531826Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:04:50.6532045Z 2025-09-07T08:04:50.6532128Z cudagraph partition due to non gpu ops 2025-09-07T08:04:50.6532331Z cudagraph partition due to non gpu ops 2025-09-07T08:04:50.7219936Z W0907 08:04:50.721490 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] torch._dynamo hit config.recompile_limit (8) 2025-09-07T08:04:50.7220839Z W0907 08:04:50.721490 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] function: 'forward' (/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py:417) 2025-09-07T08:04:50.7222602Z W0907 08:04:50.721490 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] last reason: 10/7: self.seed == 7 # context_layer, attention_probs = self.bigbird_block_sparse_attention( # transformers/models/big_bird/modeling_big_bird.py:455 in forward (HINT: torch.compile considers integer attributes of the nn.Module to be static. If you are observing recompilation, you might want to make this integer dynamic using torch._dynamo.config.allow_unspec_int_on_nn_module = True, or convert this integer into a tensor.) 2025-09-07T08:04:50.7224274Z W0907 08:04:50.721490 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] To log all recompilation reasons, use TORCH_LOGS="recompiles". 2025-09-07T08:04:50.7225114Z W0907 08:04:50.721490 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html 2025-09-07T08:04:50.7923596Z W0907 08:04:50.790784 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] torch._dynamo hit config.recompile_limit (8) 2025-09-07T08:04:50.7924504Z W0907 08:04:50.790784 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] function: 'torch_dynamo_resume_in_forward_at_1381' (/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py:1381) 2025-09-07T08:04:50.7925732Z W0907 08:04:50.790784 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] last reason: 17/7: ___check_obj_id(self._modules['output']._modules['dense']._parameters['bias'], 139801714260304) # return F.linear(input, self.weight, self.bias) # nn/modules/linear.py:134 in forward 2025-09-07T08:04:50.7926521Z W0907 08:04:50.790784 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] To log all recompilation reasons, use TORCH_LOGS="recompiles". 2025-09-07T08:04:50.7927189Z W0907 08:04:50.790784 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html 2025-09-07T08:04:50.7954331Z W0907 08:04:50.795053 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] torch._dynamo hit config.recompile_limit (8) 2025-09-07T08:04:50.7955214Z W0907 08:04:50.795053 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] function: 'torch_dynamo_resume_in_forward_at_1468' (/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py:1468) 2025-09-07T08:04:50.7956421Z W0907 08:04:50.795053 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] last reason: 18/7: ___check_obj_id(self._modules['output']._modules['dense']._parameters['bias'], 139801714261584) # return F.linear(input, self.weight, self.bias) # nn/modules/linear.py:134 in forward 2025-09-07T08:04:50.7957375Z W0907 08:04:50.795053 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] To log all recompilation reasons, use TORCH_LOGS="recompiles". 2025-09-07T08:04:50.7958177Z W0907 08:04:50.795053 39758 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html 2025-09-07T08:04:51.1372465Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:51.1376299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2010, in torch_dynamo_resume_in_forward_at_1991 2025-09-07T08:04:51.1376940Z pooler_output = self.activation(self.pooler(sequence_output[:, 0, :])) if (self.pooler is not None) else None 2025-09-07T08:04:51.1377209Z 2025-09-07T08:04:51.1377324Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:51.1377807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2010, in torch_dynamo_resume_in_forward_at_1991 2025-09-07T08:04:51.1378403Z pooler_output = self.activation(self.pooler(sequence_output[:, 0, :])) if (self.pooler is not None) else None 2025-09-07T08:04:51.1378661Z 2025-09-07T08:04:51.1378766Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:51.1379227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2010, in torch_dynamo_resume_in_forward_at_1991 2025-09-07T08:04:51.1379792Z pooler_output = self.activation(self.pooler(sequence_output[:, 0, :])) if (self.pooler is not None) else None 2025-09-07T08:04:51.1380038Z 2025-09-07T08:04:51.4087994Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:51.4088660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2318, in torch_dynamo_resume_in_forward_at_2303 2025-09-07T08:04:51.4089175Z prediction_scores = self.cls(sequence_output) 2025-09-07T08:04:51.4090109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1682, in forward 2025-09-07T08:04:51.4090647Z prediction_scores = self.predictions(sequence_output) 2025-09-07T08:04:51.4091118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1670, in forward 2025-09-07T08:04:51.4091566Z hidden_states = self.transform(hidden_states) 2025-09-07T08:04:51.4092124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1645, in forward 2025-09-07T08:04:51.4092547Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:51.4092714Z 2025-09-07T08:04:51.4092799Z cudagraph partition due to non gpu ops 2025-09-07T08:04:51.4093035Z cudagraph partition due to non gpu ops 2025-09-07T08:04:51.4093255Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7996023Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7996321Z pass 2025-09-07T08:04:52.7996629Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:54.8391189Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:04:54.8392664Z import pynvml # type: ignore[import] 2025-09-07T08:04:56.5508759Z 2025-09-07T08:04:58.5999401Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:04:58.5999717Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:04:58.5999964Z cpu eval hf_DistilBert 2025-09-07T08:04:58.9166753Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:58.9804888Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:59.0424732Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:05.0142805Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0143112Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0143326Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0143518Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0143763Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0143959Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0144160Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0144346Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0144544Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0144743Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0144947Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0145134Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0145332Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0145563Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:05.0145956Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:05.0146280Z return mod(**inputs) 2025-09-07T08:05:05.0146691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:05:05.0147113Z dlbrt_output = self.distilbert( 2025-09-07T08:05:05.0147523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:05:05.0147920Z return self.transformer( 2025-09-07T08:05:05.0148346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:05:05.0148741Z layer_outputs = layer_module( 2025-09-07T08:05:05.0149077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:05.0149428Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:05.0150069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:05:05.0150466Z sa_output = self.attention( 2025-09-07T08:05:05.0150847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 402, in forward 2025-09-07T08:05:05.0151371Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:05.0151556Z 2025-09-07T08:05:05.0151634Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0151834Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0152029Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0152222Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0152408Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0152604Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0152799Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0153001Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0153189Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0153382Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0153572Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0153762Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0153955Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0154184Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:05.0154530Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:05.0154835Z return mod(**inputs) 2025-09-07T08:05:05.0155196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:05:05.0155581Z dlbrt_output = self.distilbert( 2025-09-07T08:05:05.0155969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:05:05.0156358Z return self.transformer( 2025-09-07T08:05:05.0156733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:05:05.0157108Z layer_outputs = layer_module( 2025-09-07T08:05:05.0157440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:05.0157787Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:05.0158177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:05:05.0158554Z sa_output = self.attention( 2025-09-07T08:05:05.0158925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 402, in forward 2025-09-07T08:05:05.0159361Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:05.0159534Z 2025-09-07T08:05:05.0159615Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0159813Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0160001Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0160194Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0160388Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0160579Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0160764Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0160954Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0161144Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0161334Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0161515Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0161702Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0161891Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0162111Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:05.0162491Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:05.0162799Z return mod(**inputs) 2025-09-07T08:05:05.0163160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:05:05.0163608Z dlbrt_output = self.distilbert( 2025-09-07T08:05:05.0163992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:05:05.0164370Z return self.transformer( 2025-09-07T08:05:05.0164748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:05:05.0165135Z layer_outputs = layer_module( 2025-09-07T08:05:05.0165472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:05.0165821Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:05.0166203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:05:05.0166588Z sa_output = self.attention( 2025-09-07T08:05:05.0166965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 402, in forward 2025-09-07T08:05:05.0167403Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:05.0167573Z 2025-09-07T08:05:05.0167648Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0167853Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0168054Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0168429Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0168624Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0168825Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0169024Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0169225Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0169425Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0169611Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0169808Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0170033Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0170235Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0170457Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:05.0170799Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:05.0171094Z return mod(**inputs) 2025-09-07T08:05:05.0171462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:05:05.0171855Z dlbrt_output = self.distilbert( 2025-09-07T08:05:05.0172241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:05:05.0172625Z return self.transformer( 2025-09-07T08:05:05.0172991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:05:05.0173382Z layer_outputs = layer_module( 2025-09-07T08:05:05.0173710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:05.0174050Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:05.0174432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:05:05.0174820Z sa_output = self.attention( 2025-09-07T08:05:05.0175195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 402, in forward 2025-09-07T08:05:05.0175629Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:05.0175857Z 2025-09-07T08:05:05.0175943Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0176133Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0176326Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0176516Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0176781Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0176964Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0177155Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0177347Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0177536Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0177719Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0177909Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0178101Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0178290Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0178504Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:05.0178845Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:05.0179143Z return mod(**inputs) 2025-09-07T08:05:05.0179512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:05:05.0179897Z dlbrt_output = self.distilbert( 2025-09-07T08:05:05.0180266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:05:05.0180643Z return self.transformer( 2025-09-07T08:05:05.0181017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:05:05.0181398Z layer_outputs = layer_module( 2025-09-07T08:05:05.0181727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:05.0182063Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:05.0182455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:05:05.0182838Z sa_output = self.attention( 2025-09-07T08:05:05.0183211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 402, in forward 2025-09-07T08:05:05.0183638Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:05.0183816Z 2025-09-07T08:05:05.0183887Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0184086Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0184284Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0184479Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0184665Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0184854Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0185046Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0185240Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0185426Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0185618Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0185811Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0186007Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0186195Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0186415Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:05.0186754Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:05.0187054Z return mod(**inputs) 2025-09-07T08:05:05.0187406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:05:05.0187789Z dlbrt_output = self.distilbert( 2025-09-07T08:05:05.0188782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:05:05.0189174Z return self.transformer( 2025-09-07T08:05:05.0189553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:05:05.0189932Z layer_outputs = layer_module( 2025-09-07T08:05:05.0190308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:05.0190659Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:05.0191050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:05:05.0191434Z sa_output = self.attention( 2025-09-07T08:05:05.0191801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 402, in forward 2025-09-07T08:05:05.0192241Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:05.0192419Z 2025-09-07T08:05:05.0192493Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0192751Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0192974Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0193171Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0193360Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0193553Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0193747Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0193941Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0194127Z cudagraph partition due to non gpu ops 2025-09-07T08:05:05.0194323Z cudagraph partition due to non gpu ops 2025-09-07T08:05:13.7234021Z pass 2025-09-07T08:05:13.7234450Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:15.5415369Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:05:15.5416794Z import pynvml # type: ignore[import] 2025-09-07T08:05:17.2560048Z 2025-09-07T08:05:21.0282667Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:05:21.0283035Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:05:21.0283296Z cpu eval hf_GPT2 2025-09-07T08:05:21.2737967Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:21.3491962Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:21.4220731Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:30.7774414Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7774727Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7774968Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7775183Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7775403Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7775618Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7775818Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7776033Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7776236Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7776428Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7776620Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7776813Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7777046Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7777431Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7777761Z return mod(**inputs) 2025-09-07T08:05:30.7778375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7778782Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7779171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7779538Z outputs = block( 2025-09-07T08:05:30.7779953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7780304Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7780681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7781055Z return func(*args, **kwargs) 2025-09-07T08:05:30.7781421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7781992Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7782363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7782747Z return func(*args, **kwargs) 2025-09-07T08:05:30.7783109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.7783603Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.7784059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7784442Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7784634Z 2025-09-07T08:05:30.7784742Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7785104Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7785430Z return mod(**inputs) 2025-09-07T08:05:30.7785791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7786178Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7786560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7786926Z outputs = block( 2025-09-07T08:05:30.7787251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7787596Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7787966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7788333Z return func(*args, **kwargs) 2025-09-07T08:05:30.7788686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7789077Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7789444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7789803Z return func(*args, **kwargs) 2025-09-07T08:05:30.7790164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.7790640Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.7791095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7791476Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7791653Z 2025-09-07T08:05:30.7791756Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7792113Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7792490Z return mod(**inputs) 2025-09-07T08:05:30.7792843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7793230Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7793607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7794004Z outputs = block( 2025-09-07T08:05:30.7794333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7794684Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7795053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7795415Z return func(*args, **kwargs) 2025-09-07T08:05:30.7795776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7796165Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7796541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7796894Z return func(*args, **kwargs) 2025-09-07T08:05:30.7797252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.7797737Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.7798181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7798555Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7798726Z 2025-09-07T08:05:30.7798808Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7799026Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7799238Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7799436Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7799673Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7800025Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7800354Z return mod(**inputs) 2025-09-07T08:05:30.7800718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7801104Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7801486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7801836Z outputs = block( 2025-09-07T08:05:30.7802156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7802504Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7802885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7803241Z return func(*args, **kwargs) 2025-09-07T08:05:30.7803591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7803984Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7804355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7804711Z return func(*args, **kwargs) 2025-09-07T08:05:30.7805068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.7805475Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.7805916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:05:30.7806431Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:30.7806609Z 2025-09-07T08:05:30.7806723Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7807073Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7807416Z return mod(**inputs) 2025-09-07T08:05:30.7807765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7808147Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7808516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7808862Z outputs = block( 2025-09-07T08:05:30.7809179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7809531Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7809903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7810260Z return func(*args, **kwargs) 2025-09-07T08:05:30.7810607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7811002Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7811369Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7811725Z return func(*args, **kwargs) 2025-09-07T08:05:30.7812072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.7812460Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.7812894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:05:30.7813340Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:05:30.7813495Z 2025-09-07T08:05:30.7813603Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7813944Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7814253Z return mod(**inputs) 2025-09-07T08:05:30.7814603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7814983Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7815356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7815703Z outputs = block( 2025-09-07T08:05:30.7816015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7816370Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7816737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7817082Z return func(*args, **kwargs) 2025-09-07T08:05:30.7817437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7817821Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7818193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7818552Z return func(*args, **kwargs) 2025-09-07T08:05:30.7818903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.7819277Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.7819653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7820042Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7820210Z 2025-09-07T08:05:30.7820321Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7820665Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7822276Z return mod(**inputs) 2025-09-07T08:05:30.7822627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7823009Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7823374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7823731Z outputs = block( 2025-09-07T08:05:30.7824044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7824392Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7824755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7825113Z return func(*args, **kwargs) 2025-09-07T08:05:30.7825467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7825840Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7826216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7826557Z return func(*args, **kwargs) 2025-09-07T08:05:30.7826915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.7827286Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.7827634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7828016Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7828178Z 2025-09-07T08:05:30.7828289Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7828627Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7828946Z return mod(**inputs) 2025-09-07T08:05:30.7829294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7829673Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7830033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7830387Z outputs = block( 2025-09-07T08:05:30.7830698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7831045Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7831403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7831757Z return func(*args, **kwargs) 2025-09-07T08:05:30.7832115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.7832499Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.7832891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.7833260Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.7833599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7833970Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7834141Z 2025-09-07T08:05:30.7834281Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7834639Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7834957Z return mod(**inputs) 2025-09-07T08:05:30.7835299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7835706Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7836087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7836436Z outputs = block( 2025-09-07T08:05:30.7836751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7837101Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7837468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7837830Z return func(*args, **kwargs) 2025-09-07T08:05:30.7838180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.7838577Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.7838976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.7839354Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.7839697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7840079Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7840244Z 2025-09-07T08:05:30.7840345Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7840700Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7841017Z return mod(**inputs) 2025-09-07T08:05:30.7841362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7841735Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7842110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7842470Z outputs = block( 2025-09-07T08:05:30.7842782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7843131Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7843490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7843850Z return func(*args, **kwargs) 2025-09-07T08:05:30.7844199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.7844597Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.7844988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.7845364Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.7845707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7846081Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7846258Z 2025-09-07T08:05:30.7846362Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7846711Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7847025Z return mod(**inputs) 2025-09-07T08:05:30.7847389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7847771Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7848149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7848514Z outputs = block( 2025-09-07T08:05:30.7848855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7849193Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7849553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7849909Z return func(*args, **kwargs) 2025-09-07T08:05:30.7850257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.7850643Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.7851021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:05:30.7851388Z hidden_states = self.act(hidden_states) 2025-09-07T08:05:30.7851724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:30.7852159Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:30.7852389Z 2025-09-07T08:05:30.7852490Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7852839Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7853149Z return mod(**inputs) 2025-09-07T08:05:30.7853497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7853867Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7854236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7854594Z outputs = block( 2025-09-07T08:05:30.7854904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7855247Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7855608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7855970Z return func(*args, **kwargs) 2025-09-07T08:05:30.7856316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.7856701Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.7857077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.7857457Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.7857804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7858185Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7858345Z 2025-09-07T08:05:30.7858459Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7858798Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7859110Z return mod(**inputs) 2025-09-07T08:05:30.7859451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7859832Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7860192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7860553Z outputs = block( 2025-09-07T08:05:30.7860898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7861255Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7861616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7862008Z return func(*args, **kwargs) 2025-09-07T08:05:30.7862354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.7862748Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.7863137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.7863515Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.7863862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7864234Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7864409Z 2025-09-07T08:05:30.7864509Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7864862Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7865182Z return mod(**inputs) 2025-09-07T08:05:30.7865522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7865898Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7866270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7866628Z outputs = block( 2025-09-07T08:05:30.7866937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7867275Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7867643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7868004Z return func(*args, **kwargs) 2025-09-07T08:05:30.7868543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 442, in forward 2025-09-07T08:05:30.7868945Z hidden_states = residual + feed_forward_hidden_states 2025-09-07T08:05:30.7869100Z 2025-09-07T08:05:30.7869201Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7869551Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7869862Z return mod(**inputs) 2025-09-07T08:05:30.7870210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7870582Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7870959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7871315Z outputs = block( 2025-09-07T08:05:30.7871629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7871976Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7872333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7872689Z return func(*args, **kwargs) 2025-09-07T08:05:30.7873039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7873423Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7873786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7874139Z return func(*args, **kwargs) 2025-09-07T08:05:30.7874572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.7875052Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.7875497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7875933Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7876109Z 2025-09-07T08:05:30.7876211Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7876569Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7876884Z return mod(**inputs) 2025-09-07T08:05:30.7877236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7877611Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7877992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7878355Z outputs = block( 2025-09-07T08:05:30.7878667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7879017Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7879383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7879741Z return func(*args, **kwargs) 2025-09-07T08:05:30.7880098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7880477Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7880855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7881216Z return func(*args, **kwargs) 2025-09-07T08:05:30.7881568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.7882036Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.7882483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7882867Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7883041Z 2025-09-07T08:05:30.7883124Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7883340Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7883540Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7883746Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7883980Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7884335Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7884642Z return mod(**inputs) 2025-09-07T08:05:30.7884997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7885381Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7885760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7886118Z outputs = block( 2025-09-07T08:05:30.7886424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7886776Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7887144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7887501Z return func(*args, **kwargs) 2025-09-07T08:05:30.7887889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7888264Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7888639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7889032Z return func(*args, **kwargs) 2025-09-07T08:05:30.7889390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.7889783Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.7890209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:05:30.7890677Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:30.7890861Z 2025-09-07T08:05:30.7890961Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7891313Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7891627Z return mod(**inputs) 2025-09-07T08:05:30.7891967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7892353Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7892728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7893088Z outputs = block( 2025-09-07T08:05:30.7893393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7893750Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7894110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7894468Z return func(*args, **kwargs) 2025-09-07T08:05:30.7894827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7895205Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7895578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7895941Z return func(*args, **kwargs) 2025-09-07T08:05:30.7896300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.7896685Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.7897117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:05:30.7897560Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:05:30.7897715Z 2025-09-07T08:05:30.7897826Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7898166Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7898481Z return mod(**inputs) 2025-09-07T08:05:30.7898835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7899227Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7899607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7899958Z outputs = block( 2025-09-07T08:05:30.7900274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7900623Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7900993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7901383Z return func(*args, **kwargs) 2025-09-07T08:05:30.7901730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7902108Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7902480Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7902866Z return func(*args, **kwargs) 2025-09-07T08:05:30.7903212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.7903584Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.7903931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7904317Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7904486Z 2025-09-07T08:05:30.7904601Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7904942Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7905256Z return mod(**inputs) 2025-09-07T08:05:30.7905601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7905983Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7906347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7906702Z outputs = block( 2025-09-07T08:05:30.7907013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7907363Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7907727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7908079Z return func(*args, **kwargs) 2025-09-07T08:05:30.7908436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7908817Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7909183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7909542Z return func(*args, **kwargs) 2025-09-07T08:05:30.7909885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.7910260Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.7910604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7910988Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7911152Z 2025-09-07T08:05:30.7911268Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7911618Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7911922Z return mod(**inputs) 2025-09-07T08:05:30.7912271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7912657Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7913024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7913376Z outputs = block( 2025-09-07T08:05:30.7913682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7914026Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7914386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7914772Z return func(*args, **kwargs) 2025-09-07T08:05:30.7915130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.7915529Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.7915956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.7916331Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.7916670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7917055Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7917220Z 2025-09-07T08:05:30.7917325Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7917678Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7917995Z return mod(**inputs) 2025-09-07T08:05:30.7918345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7918721Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7919101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7919462Z outputs = block( 2025-09-07T08:05:30.7919772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7920112Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7920472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7920830Z return func(*args, **kwargs) 2025-09-07T08:05:30.7921181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.7921573Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.7921964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.7922333Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.7922682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7923063Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7923226Z 2025-09-07T08:05:30.7923334Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7923676Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7923985Z return mod(**inputs) 2025-09-07T08:05:30.7924328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7924709Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7925074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7925430Z outputs = block( 2025-09-07T08:05:30.7925742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7926087Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7926446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7926796Z return func(*args, **kwargs) 2025-09-07T08:05:30.7927149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.7927541Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.7927983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:05:30.7928358Z hidden_states = self.act(hidden_states) 2025-09-07T08:05:30.7928690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:30.7929132Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:30.7929396Z 2025-09-07T08:05:30.7929500Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7929853Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7930166Z return mod(**inputs) 2025-09-07T08:05:30.7930504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7930886Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7931264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7931630Z outputs = block( 2025-09-07T08:05:30.7931941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7932293Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7932661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7933022Z return func(*args, **kwargs) 2025-09-07T08:05:30.7933376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.7933759Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.7934152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.7934531Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.7934885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7935267Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7935429Z 2025-09-07T08:05:30.7935531Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7935886Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7936198Z return mod(**inputs) 2025-09-07T08:05:30.7936541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7936917Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7937290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7937644Z outputs = block( 2025-09-07T08:05:30.7937958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7938309Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7938669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7939035Z return func(*args, **kwargs) 2025-09-07T08:05:30.7939393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.7939790Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.7940185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.7940559Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.7940911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7941324Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7941490Z 2025-09-07T08:05:30.7941604Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7941947Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7942258Z return mod(**inputs) 2025-09-07T08:05:30.7942639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7943022Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7943398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7943750Z outputs = block( 2025-09-07T08:05:30.7944064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7944416Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7944787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7945150Z return func(*args, **kwargs) 2025-09-07T08:05:30.7945501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7945891Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7946272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7946635Z return func(*args, **kwargs) 2025-09-07T08:05:30.7946983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.7947468Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.7947915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7948304Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7948467Z 2025-09-07T08:05:30.7948577Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7948919Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7949236Z return mod(**inputs) 2025-09-07T08:05:30.7949582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7949960Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7950337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7950695Z outputs = block( 2025-09-07T08:05:30.7951009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7951357Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7951717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7952080Z return func(*args, **kwargs) 2025-09-07T08:05:30.7952440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7952825Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7953197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7953551Z return func(*args, **kwargs) 2025-09-07T08:05:30.7953906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.7954379Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.7954850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7955237Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7955401Z 2025-09-07T08:05:30.7955489Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7955695Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7955938Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7956142Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.7956373Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7956714Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7957028Z return mod(**inputs) 2025-09-07T08:05:30.7957373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7957748Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7958124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7958472Z outputs = block( 2025-09-07T08:05:30.7958783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7959130Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7959492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7959849Z return func(*args, **kwargs) 2025-09-07T08:05:30.7960197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7960575Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7960947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7961303Z return func(*args, **kwargs) 2025-09-07T08:05:30.7961652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.7962045Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.7962468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:05:30.7962931Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:30.7963105Z 2025-09-07T08:05:30.7963214Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7963556Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7963867Z return mod(**inputs) 2025-09-07T08:05:30.7964217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7964597Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7964974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7965322Z outputs = block( 2025-09-07T08:05:30.7965635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7965986Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7966347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7966700Z return func(*args, **kwargs) 2025-09-07T08:05:30.7967051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7967428Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7967799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7968186Z return func(*args, **kwargs) 2025-09-07T08:05:30.7968579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.7968969Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.7969401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:05:30.7969928Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:05:30.7970084Z 2025-09-07T08:05:30.7970194Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7970535Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7970848Z return mod(**inputs) 2025-09-07T08:05:30.7971200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7971582Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7971951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7972310Z outputs = block( 2025-09-07T08:05:30.7972629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7972981Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7973344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7973693Z return func(*args, **kwargs) 2025-09-07T08:05:30.7974045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7974422Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7974788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7975142Z return func(*args, **kwargs) 2025-09-07T08:05:30.7975487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.7975865Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.7976209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7976588Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7976753Z 2025-09-07T08:05:30.7976855Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7977201Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7977506Z return mod(**inputs) 2025-09-07T08:05:30.7977853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7978238Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7978597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7978950Z outputs = block( 2025-09-07T08:05:30.7979258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7979615Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7979975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7980323Z return func(*args, **kwargs) 2025-09-07T08:05:30.7980675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.7981053Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.7981424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7981818Z return func(*args, **kwargs) 2025-09-07T08:05:30.7982176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.7982549Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.7982921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7983302Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7983464Z 2025-09-07T08:05:30.7983566Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7983916Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7984225Z return mod(**inputs) 2025-09-07T08:05:30.7984567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7984937Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7985307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7985661Z outputs = block( 2025-09-07T08:05:30.7985970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7986319Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7986670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7987024Z return func(*args, **kwargs) 2025-09-07T08:05:30.7987374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.7987763Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.7988155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.7988519Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.7988858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7989238Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7989404Z 2025-09-07T08:05:30.7989539Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7989890Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7990197Z return mod(**inputs) 2025-09-07T08:05:30.7990538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7990918Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7991288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7991646Z outputs = block( 2025-09-07T08:05:30.7991947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7992294Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7992651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7993011Z return func(*args, **kwargs) 2025-09-07T08:05:30.7993361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.7993745Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.7994133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.7994505Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.7994890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.7995263Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.7995434Z 2025-09-07T08:05:30.7995536Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.7995883Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.7996225Z return mod(**inputs) 2025-09-07T08:05:30.7996571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.7996943Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.7997315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.7997668Z outputs = block( 2025-09-07T08:05:30.7997980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.7998332Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.7998691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.7999045Z return func(*args, **kwargs) 2025-09-07T08:05:30.7999396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.7999787Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8000170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:05:30.8000541Z hidden_states = self.act(hidden_states) 2025-09-07T08:05:30.8000878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:30.8001316Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:30.8001538Z 2025-09-07T08:05:30.8001647Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8001991Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8002309Z return mod(**inputs) 2025-09-07T08:05:30.8002662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8003041Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8003413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8003763Z outputs = block( 2025-09-07T08:05:30.8004074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8004421Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8004789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8005140Z return func(*args, **kwargs) 2025-09-07T08:05:30.8005495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8005885Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8006280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8006658Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8007000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8007385Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8007556Z 2025-09-07T08:05:30.8007656Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8008032Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8008349Z return mod(**inputs) 2025-09-07T08:05:30.8008687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8009069Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8009467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8009820Z outputs = block( 2025-09-07T08:05:30.8010124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8010473Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8010837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8011192Z return func(*args, **kwargs) 2025-09-07T08:05:30.8011548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8011933Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8012317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8012696Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8013042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8013424Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8013590Z 2025-09-07T08:05:30.8013692Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8039025Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8039374Z return mod(**inputs) 2025-09-07T08:05:30.8039780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8040197Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8040585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8040961Z outputs = block( 2025-09-07T08:05:30.8041282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8041629Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8041996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8042350Z return func(*args, **kwargs) 2025-09-07T08:05:30.8042702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 442, in forward 2025-09-07T08:05:30.8043098Z hidden_states = residual + feed_forward_hidden_states 2025-09-07T08:05:30.8043251Z 2025-09-07T08:05:30.8043363Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8043707Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8044019Z return mod(**inputs) 2025-09-07T08:05:30.8044361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8044752Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8045126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8045478Z outputs = block( 2025-09-07T08:05:30.8045799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8046146Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8046603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8046973Z return func(*args, **kwargs) 2025-09-07T08:05:30.8047318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8047709Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8048144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8048492Z return func(*args, **kwargs) 2025-09-07T08:05:30.8048836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8049313Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8049770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8050158Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8050326Z 2025-09-07T08:05:30.8050439Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8050785Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8051100Z return mod(**inputs) 2025-09-07T08:05:30.8051441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8051822Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8052194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8052540Z outputs = block( 2025-09-07T08:05:30.8052845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8053193Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8053559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8053907Z return func(*args, **kwargs) 2025-09-07T08:05:30.8054253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8054632Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8055001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8055352Z return func(*args, **kwargs) 2025-09-07T08:05:30.8055692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8056158Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8056600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8056987Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8057149Z 2025-09-07T08:05:30.8057238Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8057442Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8057646Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8057855Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8058080Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8058422Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8058739Z return mod(**inputs) 2025-09-07T08:05:30.8059087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8059462Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8059886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8060239Z outputs = block( 2025-09-07T08:05:30.8060552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8060902Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8061299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8061647Z return func(*args, **kwargs) 2025-09-07T08:05:30.8062010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8062385Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8062759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8063112Z return func(*args, **kwargs) 2025-09-07T08:05:30.8063461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8063851Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8064276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:05:30.8064740Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:30.8064916Z 2025-09-07T08:05:30.8065022Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8065360Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8065668Z return mod(**inputs) 2025-09-07T08:05:30.8066011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8066389Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8066753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8067106Z outputs = block( 2025-09-07T08:05:30.8067413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8067764Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8068126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8068528Z return func(*args, **kwargs) 2025-09-07T08:05:30.8068886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8069256Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8069620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8069962Z return func(*args, **kwargs) 2025-09-07T08:05:30.8070327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8070716Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8071142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:05:30.8071588Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:05:30.8071744Z 2025-09-07T08:05:30.8071845Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8072192Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8072503Z return mod(**inputs) 2025-09-07T08:05:30.8072842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8073221Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8073650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8073999Z outputs = block( 2025-09-07T08:05:30.8074307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8074692Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8075043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8075389Z return func(*args, **kwargs) 2025-09-07T08:05:30.8075729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8076107Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8076469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8076812Z return func(*args, **kwargs) 2025-09-07T08:05:30.8077158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8077525Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8077869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8078250Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8078417Z 2025-09-07T08:05:30.8078518Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8078860Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8079166Z return mod(**inputs) 2025-09-07T08:05:30.8079509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8079872Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8080244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8080591Z outputs = block( 2025-09-07T08:05:30.8080893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8081237Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8081583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8081929Z return func(*args, **kwargs) 2025-09-07T08:05:30.8082276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8082645Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8083002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8083352Z return func(*args, **kwargs) 2025-09-07T08:05:30.8083701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8084067Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8084404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8084774Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8084942Z 2025-09-07T08:05:30.8085042Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8085386Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8085691Z return mod(**inputs) 2025-09-07T08:05:30.8086028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8086392Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8086786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8087133Z outputs = block( 2025-09-07T08:05:30.8087438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8087808Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8088163Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8088508Z return func(*args, **kwargs) 2025-09-07T08:05:30.8088852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8089235Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8089609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8089973Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8090304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8090675Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8090837Z 2025-09-07T08:05:30.8090944Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8091273Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8091575Z return mod(**inputs) 2025-09-07T08:05:30.8091912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8092287Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8092641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8092984Z outputs = block( 2025-09-07T08:05:30.8093290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8093630Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8093987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8094331Z return func(*args, **kwargs) 2025-09-07T08:05:30.8094681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8095064Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8095448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8095815Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8096149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8096524Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8096690Z 2025-09-07T08:05:30.8096789Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8097126Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8097424Z return mod(**inputs) 2025-09-07T08:05:30.8097763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8098135Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8098502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8098855Z outputs = block( 2025-09-07T08:05:30.8099154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8099524Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8099879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8100222Z return func(*args, **kwargs) 2025-09-07T08:05:30.8100557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8100968Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8101347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:05:30.8101716Z hidden_states = self.act(hidden_states) 2025-09-07T08:05:30.8102047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:30.8102465Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:30.8102694Z 2025-09-07T08:05:30.8102795Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8103134Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8103437Z return mod(**inputs) 2025-09-07T08:05:30.8103775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8104143Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8104539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8104886Z outputs = block( 2025-09-07T08:05:30.8105189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8105522Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8105867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8106214Z return func(*args, **kwargs) 2025-09-07T08:05:30.8106555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8106933Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8107311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8107674Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8108010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8108384Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8108541Z 2025-09-07T08:05:30.8108649Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8108976Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8109279Z return mod(**inputs) 2025-09-07T08:05:30.8109611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8109977Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8110333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8110676Z outputs = block( 2025-09-07T08:05:30.8110980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8111319Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8111673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8112013Z return func(*args, **kwargs) 2025-09-07T08:05:30.8112397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8112786Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8113173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8113567Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8113897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8114268Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8114435Z 2025-09-07T08:05:30.8114533Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8114870Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8115167Z return mod(**inputs) 2025-09-07T08:05:30.8115506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8115875Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8116242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8116585Z outputs = block( 2025-09-07T08:05:30.8116886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8117237Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8117595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8117939Z return func(*args, **kwargs) 2025-09-07T08:05:30.8118290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8118666Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8119037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8119388Z return func(*args, **kwargs) 2025-09-07T08:05:30.8119730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8120201Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8120640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8121013Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8121174Z 2025-09-07T08:05:30.8121280Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8121614Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8121920Z return mod(**inputs) 2025-09-07T08:05:30.8122265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8122640Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8123000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8123360Z outputs = block( 2025-09-07T08:05:30.8123662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8123996Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8124350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8124686Z return func(*args, **kwargs) 2025-09-07T08:05:30.8125034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8125407Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8125805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8126157Z return func(*args, **kwargs) 2025-09-07T08:05:30.8126503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8126995Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8127430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8127801Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8127962Z 2025-09-07T08:05:30.8128041Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8128251Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8128456Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8128921Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8129150Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8129482Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8129791Z return mod(**inputs) 2025-09-07T08:05:30.8130136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8130513Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8130871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8131216Z outputs = block( 2025-09-07T08:05:30.8131521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8131857Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8132211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8132547Z return func(*args, **kwargs) 2025-09-07T08:05:30.8132890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8133259Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8133624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8133966Z return func(*args, **kwargs) 2025-09-07T08:05:30.8134303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8134681Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8135097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:05:30.8135550Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:30.8135719Z 2025-09-07T08:05:30.8135815Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8136154Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8136459Z return mod(**inputs) 2025-09-07T08:05:30.8136800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8137166Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8137522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8137867Z outputs = block( 2025-09-07T08:05:30.8138167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8138503Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8138884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8139234Z return func(*args, **kwargs) 2025-09-07T08:05:30.8139586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8139987Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8140351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8140688Z return func(*args, **kwargs) 2025-09-07T08:05:30.8141036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8141413Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8141829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:05:30.8142264Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:05:30.8142420Z 2025-09-07T08:05:30.8142521Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8142864Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8143175Z return mod(**inputs) 2025-09-07T08:05:30.8143513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8143885Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8144246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8144588Z outputs = block( 2025-09-07T08:05:30.8144891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8145226Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8145576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8145920Z return func(*args, **kwargs) 2025-09-07T08:05:30.8146264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8146636Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8146994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8147328Z return func(*args, **kwargs) 2025-09-07T08:05:30.8147668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8148030Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8148363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8148729Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8148897Z 2025-09-07T08:05:30.8148997Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8149339Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8149643Z return mod(**inputs) 2025-09-07T08:05:30.8149980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8150342Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8150703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8151046Z outputs = block( 2025-09-07T08:05:30.8151344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8151683Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8152059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8152402Z return func(*args, **kwargs) 2025-09-07T08:05:30.8152741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8153149Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8153502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8153838Z return func(*args, **kwargs) 2025-09-07T08:05:30.8154175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8154530Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8154864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8155228Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8155396Z 2025-09-07T08:05:30.8155492Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8155825Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8156134Z return mod(**inputs) 2025-09-07T08:05:30.8156461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8156823Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8157180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8157523Z outputs = block( 2025-09-07T08:05:30.8157824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8158152Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8158510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8158853Z return func(*args, **kwargs) 2025-09-07T08:05:30.8159194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8159578Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8159953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8160311Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8160640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8161008Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8161167Z 2025-09-07T08:05:30.8161264Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8161602Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8161904Z return mod(**inputs) 2025-09-07T08:05:30.8162239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8162319Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8162554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8162613Z outputs = block( 2025-09-07T08:05:30.8162818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8162898Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8163119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8163191Z return func(*args, **kwargs) 2025-09-07T08:05:30.8163448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8163552Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8163776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8163881Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8164084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8164191Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8164194Z 2025-09-07T08:05:30.8164299Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8164481Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8164541Z return mod(**inputs) 2025-09-07T08:05:30.8164778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8164855Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8165086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8165148Z outputs = block( 2025-09-07T08:05:30.8165351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8165430Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8165652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8165725Z return func(*args, **kwargs) 2025-09-07T08:05:30.8165949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8166052Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8166275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:05:30.8166346Z hidden_states = self.act(hidden_states) 2025-09-07T08:05:30.8166549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:30.8166717Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:30.8166720Z 2025-09-07T08:05:30.8166825Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8167009Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8167071Z return mod(**inputs) 2025-09-07T08:05:30.8167308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8167389Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8167622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8167681Z outputs = block( 2025-09-07T08:05:30.8167890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8167967Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8168196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8168324Z return func(*args, **kwargs) 2025-09-07T08:05:30.8168560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8168664Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8168953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8169038Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8169301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8169412Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8169465Z 2025-09-07T08:05:30.8169575Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8169761Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8169823Z return mod(**inputs) 2025-09-07T08:05:30.8170064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8170142Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8170381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8170443Z outputs = block( 2025-09-07T08:05:30.8170658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8170732Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8170963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8171037Z return func(*args, **kwargs) 2025-09-07T08:05:30.8171271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8171372Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8171598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8171680Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8171893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8172004Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8172007Z 2025-09-07T08:05:30.8172112Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8172299Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8172370Z return mod(**inputs) 2025-09-07T08:05:30.8172603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8172683Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8172921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8172980Z outputs = block( 2025-09-07T08:05:30.8173196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8173269Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8173494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8173568Z return func(*args, **kwargs) 2025-09-07T08:05:30.8173799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 442, in forward 2025-09-07T08:05:30.8173909Z hidden_states = residual + feed_forward_hidden_states 2025-09-07T08:05:30.8173913Z 2025-09-07T08:05:30.8174008Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8174190Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8174263Z return mod(**inputs) 2025-09-07T08:05:30.8174493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8174604Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8174832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8174899Z outputs = block( 2025-09-07T08:05:30.8175105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8175208Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8175434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8175499Z return func(*args, **kwargs) 2025-09-07T08:05:30.8175734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8175816Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8176040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8176112Z return func(*args, **kwargs) 2025-09-07T08:05:30.8176337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8176519Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8176722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8176839Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8176842Z 2025-09-07T08:05:30.8176937Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8177119Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8177187Z return mod(**inputs) 2025-09-07T08:05:30.8177419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8177505Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8177728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8177786Z outputs = block( 2025-09-07T08:05:30.8177998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8178071Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8178299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8178363Z return func(*args, **kwargs) 2025-09-07T08:05:30.8178584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8178672Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8178894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8178966Z return func(*args, **kwargs) 2025-09-07T08:05:30.8179191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8179372Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8179572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8179679Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8179682Z 2025-09-07T08:05:30.8179767Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8179843Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8179921Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8179993Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8180119Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8180310Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8180371Z return mod(**inputs) 2025-09-07T08:05:30.8180610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8180724Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8180947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8181013Z outputs = block( 2025-09-07T08:05:30.8181219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8181297Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8181520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8181586Z return func(*args, **kwargs) 2025-09-07T08:05:30.8181817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8181900Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8182130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8182193Z return func(*args, **kwargs) 2025-09-07T08:05:30.8182426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8182518Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8182787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:05:30.8182916Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:30.8182919Z 2025-09-07T08:05:30.8183017Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8183207Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8183269Z return mod(**inputs) 2025-09-07T08:05:30.8183500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8183588Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8183813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8183881Z outputs = block( 2025-09-07T08:05:30.8184089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8184170Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8184395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8184459Z return func(*args, **kwargs) 2025-09-07T08:05:30.8184692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8184774Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8185003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8185065Z return func(*args, **kwargs) 2025-09-07T08:05:30.8185289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8185387Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8185655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:05:30.8185769Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:05:30.8185797Z 2025-09-07T08:05:30.8185898Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8186085Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8186147Z return mod(**inputs) 2025-09-07T08:05:30.8186402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8186489Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8186714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8186781Z outputs = block( 2025-09-07T08:05:30.8186988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8187062Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8187294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8187359Z return func(*args, **kwargs) 2025-09-07T08:05:30.8187594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8187676Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8187896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8187968Z return func(*args, **kwargs) 2025-09-07T08:05:30.8188195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8188281Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8188482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8188603Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8188606Z 2025-09-07T08:05:30.8188703Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8188884Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8188955Z return mod(**inputs) 2025-09-07T08:05:30.8189188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8189273Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8189496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8189556Z outputs = block( 2025-09-07T08:05:30.8189765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8189838Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8190068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8190134Z return func(*args, **kwargs) 2025-09-07T08:05:30.8190359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8190448Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8190667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8190736Z return func(*args, **kwargs) 2025-09-07T08:05:30.8190962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8191048Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8191250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8191391Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8191395Z 2025-09-07T08:05:30.8191501Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8191682Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8191750Z return mod(**inputs) 2025-09-07T08:05:30.8192012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8192092Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8192326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8192387Z outputs = block( 2025-09-07T08:05:30.8192599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8192673Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8192899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8192970Z return func(*args, **kwargs) 2025-09-07T08:05:30.8193200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8193306Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8193534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8193615Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8193815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8193920Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8193923Z 2025-09-07T08:05:30.8194029Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8194212Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8194282Z return mod(**inputs) 2025-09-07T08:05:30.8194513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8194592Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8194827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8194885Z outputs = block( 2025-09-07T08:05:30.8195095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8195168Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8195391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8195462Z return func(*args, **kwargs) 2025-09-07T08:05:30.8195690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8195796Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8196024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8196107Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8196309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8196416Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8196419Z 2025-09-07T08:05:30.8196525Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8196709Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8196777Z return mod(**inputs) 2025-09-07T08:05:30.8197033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8197114Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8197350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8197442Z outputs = block( 2025-09-07T08:05:30.8197657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8197731Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8197963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8198027Z return func(*args, **kwargs) 2025-09-07T08:05:30.8198255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8198356Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8198585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:05:30.8198667Z hidden_states = self.act(hidden_states) 2025-09-07T08:05:30.8198863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:30.8199035Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:30.8199037Z 2025-09-07T08:05:30.8199144Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8199328Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8199400Z return mod(**inputs) 2025-09-07T08:05:30.8199633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8199719Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8199950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8200009Z outputs = block( 2025-09-07T08:05:30.8200221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8200297Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8200534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8200597Z return func(*args, **kwargs) 2025-09-07T08:05:30.8200822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8200924Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8201152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8201239Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8201440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8201548Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8201562Z 2025-09-07T08:05:30.8201658Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8201844Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8201912Z return mod(**inputs) 2025-09-07T08:05:30.8202141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8202224Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8202450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8202548Z outputs = block( 2025-09-07T08:05:30.8202758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8202830Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8203056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8203148Z return func(*args, **kwargs) 2025-09-07T08:05:30.8203380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8203482Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8203705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8203795Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8204003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8204110Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8204119Z 2025-09-07T08:05:30.8204215Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8204397Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8204470Z return mod(**inputs) 2025-09-07T08:05:30.8204700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8204785Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8205012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8205069Z outputs = block( 2025-09-07T08:05:30.8205282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8205357Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8205587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8205651Z return func(*args, **kwargs) 2025-09-07T08:05:30.8205876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8205968Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8206193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8206265Z return func(*args, **kwargs) 2025-09-07T08:05:30.8206492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8206671Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8206878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8206985Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8206988Z 2025-09-07T08:05:30.8207090Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8207273Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8207340Z return mod(**inputs) 2025-09-07T08:05:30.8207572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8207650Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8207883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8207941Z outputs = block( 2025-09-07T08:05:30.8208179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8208251Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8208472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8208542Z return func(*args, **kwargs) 2025-09-07T08:05:30.8208798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8208887Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8209107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8209176Z return func(*args, **kwargs) 2025-09-07T08:05:30.8209402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8209574Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8209783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8209890Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8209893Z 2025-09-07T08:05:30.8209976Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8210047Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8210119Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8210195Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8210290Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8210481Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8210541Z return mod(**inputs) 2025-09-07T08:05:30.8210772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8210861Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8211086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8211153Z outputs = block( 2025-09-07T08:05:30.8211356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8211438Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8211660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8211723Z return func(*args, **kwargs) 2025-09-07T08:05:30.8211959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8212041Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8212272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8212335Z return func(*args, **kwargs) 2025-09-07T08:05:30.8212563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8212663Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8212938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:05:30.8213065Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:30.8213068Z 2025-09-07T08:05:30.8213165Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8213359Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8213420Z return mod(**inputs) 2025-09-07T08:05:30.8213650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8213766Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8213994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8214062Z outputs = block( 2025-09-07T08:05:30.8214298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8214370Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8214598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8214662Z return func(*args, **kwargs) 2025-09-07T08:05:30.8214894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8214974Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8215196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8215264Z return func(*args, **kwargs) 2025-09-07T08:05:30.8215489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8215588Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8215853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:05:30.8215964Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:05:30.8215967Z 2025-09-07T08:05:30.8216065Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8216246Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8216315Z return mod(**inputs) 2025-09-07T08:05:30.8216551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8216636Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8216864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8216922Z outputs = block( 2025-09-07T08:05:30.8217134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8217208Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8217437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8217500Z return func(*args, **kwargs) 2025-09-07T08:05:30.8217726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8217814Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8218036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8218106Z return func(*args, **kwargs) 2025-09-07T08:05:30.8218331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8218417Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8218620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8218727Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8218730Z 2025-09-07T08:05:30.8218836Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8219018Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8219087Z return mod(**inputs) 2025-09-07T08:05:30.8219346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8219422Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8219657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8219746Z outputs = block( 2025-09-07T08:05:30.8219961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8220033Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8220254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8220326Z return func(*args, **kwargs) 2025-09-07T08:05:30.8220555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8220640Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8220866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8220934Z return func(*args, **kwargs) 2025-09-07T08:05:30.8221162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8221239Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8221449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8221557Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8221560Z 2025-09-07T08:05:30.8221664Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8221845Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8221906Z return mod(**inputs) 2025-09-07T08:05:30.8222147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8222223Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8222455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8222515Z outputs = block( 2025-09-07T08:05:30.8222726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8222799Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8223020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8223092Z return func(*args, **kwargs) 2025-09-07T08:05:30.8223319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8223419Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8223651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8223724Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8223935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8224045Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8224047Z 2025-09-07T08:05:30.8224148Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8224330Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8224392Z return mod(**inputs) 2025-09-07T08:05:30.8224636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8224713Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8224991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8225050Z outputs = block( 2025-09-07T08:05:30.8225261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8225363Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8225587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8225664Z return func(*args, **kwargs) 2025-09-07T08:05:30.8225892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8226000Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8226229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8226308Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8226519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8226630Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8226633Z 2025-09-07T08:05:30.8226747Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8226933Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8227008Z return mod(**inputs) 2025-09-07T08:05:30.8227243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8227324Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8227563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8227624Z outputs = block( 2025-09-07T08:05:30.8227842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8227916Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8228141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8228221Z return func(*args, **kwargs) 2025-09-07T08:05:30.8228451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8228555Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8228784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:05:30.8228860Z hidden_states = self.act(hidden_states) 2025-09-07T08:05:30.8229066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:30.8229237Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:30.8229240Z 2025-09-07T08:05:30.8229350Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8229539Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8229613Z return mod(**inputs) 2025-09-07T08:05:30.8229849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8229929Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8230170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8230233Z outputs = block( 2025-09-07T08:05:30.8230445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8230550Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8230770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8230843Z return func(*args, **kwargs) 2025-09-07T08:05:30.8231067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8231194Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8231420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8231500Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8231706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8231812Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8231816Z 2025-09-07T08:05:30.8231920Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8232101Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8232170Z return mod(**inputs) 2025-09-07T08:05:30.8232397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8232476Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8232713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8232770Z outputs = block( 2025-09-07T08:05:30.8232978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8233049Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8233273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8233350Z return func(*args, **kwargs) 2025-09-07T08:05:30.8233579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8233677Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8233903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8233989Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8234187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8234292Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8234295Z 2025-09-07T08:05:30.8234398Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8234578Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8234644Z return mod(**inputs) 2025-09-07T08:05:30.8234874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8234950Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8235185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8235247Z outputs = block( 2025-09-07T08:05:30.8235455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8235526Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8235746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8235819Z return func(*args, **kwargs) 2025-09-07T08:05:30.8236043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 442, in forward 2025-09-07T08:05:30.8236178Z hidden_states = residual + feed_forward_hidden_states 2025-09-07T08:05:30.8236181Z 2025-09-07T08:05:30.8236279Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8236468Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8236560Z return mod(**inputs) 2025-09-07T08:05:30.8236796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8236878Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8237109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8237173Z outputs = block( 2025-09-07T08:05:30.8237380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8237450Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8237685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8237751Z return func(*args, **kwargs) 2025-09-07T08:05:30.8237988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8238070Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8238295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8238368Z return func(*args, **kwargs) 2025-09-07T08:05:30.8238599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8238777Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8238986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8239101Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8239104Z 2025-09-07T08:05:30.8239201Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8239386Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8239456Z return mod(**inputs) 2025-09-07T08:05:30.8239693Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8239778Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8240010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8240070Z outputs = block( 2025-09-07T08:05:30.8240288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8240363Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8240596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8240659Z return func(*args, **kwargs) 2025-09-07T08:05:30.8240903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8240983Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8241208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8241280Z return func(*args, **kwargs) 2025-09-07T08:05:30.8241512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8241691Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8241926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8242035Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8242046Z 2025-09-07T08:05:30.8242123Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8242234Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8242316Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8242385Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8242478Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8242670Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8242731Z return mod(**inputs) 2025-09-07T08:05:30.8242968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8243044Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8243280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8243340Z outputs = block( 2025-09-07T08:05:30.8243542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8243629Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8243851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8243921Z return func(*args, **kwargs) 2025-09-07T08:05:30.8244147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8244228Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8244452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8244518Z return func(*args, **kwargs) 2025-09-07T08:05:30.8244751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8244842Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8245109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:05:30.8245238Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:30.8245241Z 2025-09-07T08:05:30.8245338Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8245525Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8245587Z return mod(**inputs) 2025-09-07T08:05:30.8245827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8245906Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8246132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8246196Z outputs = block( 2025-09-07T08:05:30.8246399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8246481Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8246701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8246765Z return func(*args, **kwargs) 2025-09-07T08:05:30.8246996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8247077Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8247337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8247401Z return func(*args, **kwargs) 2025-09-07T08:05:30.8247627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8247725Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8248022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:05:30.8248133Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:05:30.8248136Z 2025-09-07T08:05:30.8248231Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8248420Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8248482Z return mod(**inputs) 2025-09-07T08:05:30.8248715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8248803Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8249031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8249095Z outputs = block( 2025-09-07T08:05:30.8249298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8249372Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8249600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8249663Z return func(*args, **kwargs) 2025-09-07T08:05:30.8249894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8249972Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8250200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8250264Z return func(*args, **kwargs) 2025-09-07T08:05:30.8250490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8250574Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8250778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8250891Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8250894Z 2025-09-07T08:05:30.8250988Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8251168Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8251235Z return mod(**inputs) 2025-09-07T08:05:30.8251463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8251550Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8251775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8251834Z outputs = block( 2025-09-07T08:05:30.8252047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8252121Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8252349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8252414Z return func(*args, **kwargs) 2025-09-07T08:05:30.8252647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8252725Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8252969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8253042Z return func(*args, **kwargs) 2025-09-07T08:05:30.8253268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8253381Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8253582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8253690Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8253693Z 2025-09-07T08:05:30.8253797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8253978Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8254052Z return mod(**inputs) 2025-09-07T08:05:30.8254287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8254373Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8254600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8254659Z outputs = block( 2025-09-07T08:05:30.8254873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8254945Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8255171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8255234Z return func(*args, **kwargs) 2025-09-07T08:05:30.8255459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8255561Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8255790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8255873Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8256072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8256181Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8256190Z 2025-09-07T08:05:30.8256285Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8256468Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8256536Z return mod(**inputs) 2025-09-07T08:05:30.8256766Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8256848Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8257079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8257138Z outputs = block( 2025-09-07T08:05:30.8257351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8257425Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8257657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8257722Z return func(*args, **kwargs) 2025-09-07T08:05:30.8257944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8258048Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8258275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8258353Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8258585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8258695Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8258705Z 2025-09-07T08:05:30.8258800Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8259014Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8259084Z return mod(**inputs) 2025-09-07T08:05:30.8259313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8259396Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8259623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8259679Z outputs = block( 2025-09-07T08:05:30.8259895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8259968Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8260201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8260266Z return func(*args, **kwargs) 2025-09-07T08:05:30.8260496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8260599Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8260825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:05:30.8260905Z hidden_states = self.act(hidden_states) 2025-09-07T08:05:30.8261099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:30.8261274Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:30.8261278Z 2025-09-07T08:05:30.8261376Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8261561Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8261634Z return mod(**inputs) 2025-09-07T08:05:30.8261865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8261953Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8262183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8262239Z outputs = block( 2025-09-07T08:05:30.8262451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8262523Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8262757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8262819Z return func(*args, **kwargs) 2025-09-07T08:05:30.8263048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8263148Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8263374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8263458Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8263658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8263773Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8263776Z 2025-09-07T08:05:30.8263873Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8264083Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8264152Z return mod(**inputs) 2025-09-07T08:05:30.8264383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8264501Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8264725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8264783Z outputs = block( 2025-09-07T08:05:30.8264995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8265066Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8265294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8265355Z return func(*args, **kwargs) 2025-09-07T08:05:30.8265590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8265683Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8265905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8265995Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8266194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8266310Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8266313Z 2025-09-07T08:05:30.8266409Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8266594Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8266661Z return mod(**inputs) 2025-09-07T08:05:30.8266891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8266977Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8267204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8267265Z outputs = block( 2025-09-07T08:05:30.8267476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8267549Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8267778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8267841Z return func(*args, **kwargs) 2025-09-07T08:05:30.8268076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8268161Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8268441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8268514Z return func(*args, **kwargs) 2025-09-07T08:05:30.8268745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8268935Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8269152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8269260Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8269272Z 2025-09-07T08:05:30.8269368Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8269552Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8269675Z return mod(**inputs) 2025-09-07T08:05:30.8269906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8269993Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8270267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8270370Z outputs = block( 2025-09-07T08:05:30.8270593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8270668Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8270910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8270976Z return func(*args, **kwargs) 2025-09-07T08:05:30.8271221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8271312Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8271548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8271619Z return func(*args, **kwargs) 2025-09-07T08:05:30.8271864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8272038Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8272260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8272371Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8272374Z 2025-09-07T08:05:30.8272466Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8272543Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8272624Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8272698Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8272797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8272995Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8273060Z return mod(**inputs) 2025-09-07T08:05:30.8273312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8273390Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8273632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8273699Z outputs = block( 2025-09-07T08:05:30.8273913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8273995Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8274230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8274294Z return func(*args, **kwargs) 2025-09-07T08:05:30.8274539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8274623Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8274862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8274926Z return func(*args, **kwargs) 2025-09-07T08:05:30.8275162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8275261Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8275572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:05:30.8275704Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:30.8275707Z 2025-09-07T08:05:30.8275807Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8275999Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8276097Z return mod(**inputs) 2025-09-07T08:05:30.8276341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8276431Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8276666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8276732Z outputs = block( 2025-09-07T08:05:30.8276947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8277026Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8277270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8277337Z return func(*args, **kwargs) 2025-09-07T08:05:30.8277585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8277671Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8277913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8277978Z return func(*args, **kwargs) 2025-09-07T08:05:30.8278215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8278315Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8278603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:05:30.8278718Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:05:30.8278721Z 2025-09-07T08:05:30.8278821Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8279012Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8279083Z return mod(**inputs) 2025-09-07T08:05:30.8279328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8279413Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8279652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8279711Z outputs = block( 2025-09-07T08:05:30.8279935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8280012Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8280256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8280323Z return func(*args, **kwargs) 2025-09-07T08:05:30.8280572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8280653Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8280886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8280960Z return func(*args, **kwargs) 2025-09-07T08:05:30.8281197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8281284Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8281525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8281638Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8281641Z 2025-09-07T08:05:30.8281749Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8281977Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8282047Z return mod(**inputs) 2025-09-07T08:05:30.8282280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8282366Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8282591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8282652Z outputs = block( 2025-09-07T08:05:30.8282862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8282937Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8283165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8283229Z return func(*args, **kwargs) 2025-09-07T08:05:30.8283458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8283546Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8283772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8283842Z return func(*args, **kwargs) 2025-09-07T08:05:30.8284071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8284145Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8284353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8284461Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8284464Z 2025-09-07T08:05:30.8284569Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8284756Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8284823Z return mod(**inputs) 2025-09-07T08:05:30.8285052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8285130Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8285364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8285423Z outputs = block( 2025-09-07T08:05:30.8285638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8285712Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8285933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8286003Z return func(*args, **kwargs) 2025-09-07T08:05:30.8286233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8286335Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8286561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8286635Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8286844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8286952Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8286989Z 2025-09-07T08:05:30.8287097Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8287282Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8287354Z return mod(**inputs) 2025-09-07T08:05:30.8287616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8287691Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8287923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8287983Z outputs = block( 2025-09-07T08:05:30.8288192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8288264Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8288490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8288563Z return func(*args, **kwargs) 2025-09-07T08:05:30.8288788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8288890Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8289119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8289201Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8289401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8289510Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8289513Z 2025-09-07T08:05:30.8289617Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8289803Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8289872Z return mod(**inputs) 2025-09-07T08:05:30.8290102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8290180Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8290422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8290480Z outputs = block( 2025-09-07T08:05:30.8290692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8290765Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8290988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8291060Z return func(*args, **kwargs) 2025-09-07T08:05:30.8291289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8291391Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8291618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:05:30.8291705Z hidden_states = self.act(hidden_states) 2025-09-07T08:05:30.8291901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:30.8292069Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:30.8292072Z 2025-09-07T08:05:30.8292178Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8292360Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8292429Z return mod(**inputs) 2025-09-07T08:05:30.8292690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8292769Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8293005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8293111Z outputs = block( 2025-09-07T08:05:30.8293322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8293395Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8293623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8293688Z return func(*args, **kwargs) 2025-09-07T08:05:30.8293916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8294019Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8294247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8294336Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8294536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8294650Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8294652Z 2025-09-07T08:05:30.8294762Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8294947Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8295018Z return mod(**inputs) 2025-09-07T08:05:30.8295251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8295330Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8295568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8295628Z outputs = block( 2025-09-07T08:05:30.8295837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8295914Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8296146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8296212Z return func(*args, **kwargs) 2025-09-07T08:05:30.8296438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8296544Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8296771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8296863Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8297063Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8297170Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8297175Z 2025-09-07T08:05:30.8297281Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8297465Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8297533Z return mod(**inputs) 2025-09-07T08:05:30.8297761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8297848Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8298071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8298131Z outputs = block( 2025-09-07T08:05:30.8298367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8298443Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8298674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8298767Z return func(*args, **kwargs) 2025-09-07T08:05:30.8298998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 442, in forward 2025-09-07T08:05:30.8299105Z hidden_states = residual + feed_forward_hidden_states 2025-09-07T08:05:30.8299108Z 2025-09-07T08:05:30.8299204Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8299396Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8299455Z return mod(**inputs) 2025-09-07T08:05:30.8299691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8299776Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8300005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8300077Z outputs = block( 2025-09-07T08:05:30.8300284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8300367Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8300595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8300658Z return func(*args, **kwargs) 2025-09-07T08:05:30.8300896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8300976Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8301211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8301274Z return func(*args, **kwargs) 2025-09-07T08:05:30.8301502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8301682Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8301888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8302001Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8302004Z 2025-09-07T08:05:30.8302100Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8302292Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8302353Z return mod(**inputs) 2025-09-07T08:05:30.8302586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8302673Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8302901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8302968Z outputs = block( 2025-09-07T08:05:30.8303173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8303245Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8303478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8303542Z return func(*args, **kwargs) 2025-09-07T08:05:30.8303774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8303883Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8304105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8304175Z return func(*args, **kwargs) 2025-09-07T08:05:30.8304399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8304608Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8304810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8304924Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8304927Z 2025-09-07T08:05:30.8305003Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8305076Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8305153Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8305228Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8305331Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8305516Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8305576Z return mod(**inputs) 2025-09-07T08:05:30.8305822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8305897Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8306129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8306189Z outputs = block( 2025-09-07T08:05:30.8306395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8306479Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8306704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8306775Z return func(*args, **kwargs) 2025-09-07T08:05:30.8307003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8307089Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8307320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8307383Z return func(*args, **kwargs) 2025-09-07T08:05:30.8307618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8307710Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8307990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:05:30.8308114Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:30.8308117Z 2025-09-07T08:05:30.8308212Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8308409Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8308474Z return mod(**inputs) 2025-09-07T08:05:30.8308714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8308792Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8309018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8309085Z outputs = block( 2025-09-07T08:05:30.8309287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8309368Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8309624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8309696Z return func(*args, **kwargs) 2025-09-07T08:05:30.8309926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8310036Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8310265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8310329Z return func(*args, **kwargs) 2025-09-07T08:05:30.8310563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8310652Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8310921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:05:30.8311033Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:05:30.8311036Z 2025-09-07T08:05:30.8311133Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8311324Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8311388Z return mod(**inputs) 2025-09-07T08:05:30.8311620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8311706Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8311934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8312000Z outputs = block( 2025-09-07T08:05:30.8312204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8312290Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8312512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8312576Z return func(*args, **kwargs) 2025-09-07T08:05:30.8312813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8312897Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8313129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8313191Z return func(*args, **kwargs) 2025-09-07T08:05:30.8313416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8313502Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8313703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8313819Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8313822Z 2025-09-07T08:05:30.8313920Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8314109Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8314173Z return mod(**inputs) 2025-09-07T08:05:30.8314403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8314485Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8314711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8314778Z outputs = block( 2025-09-07T08:05:30.8314980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8315084Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8315313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8315379Z return func(*args, **kwargs) 2025-09-07T08:05:30.8315613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8315720Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8315943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8316013Z return func(*args, **kwargs) 2025-09-07T08:05:30.8316237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8316319Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8316524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8316640Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8316642Z 2025-09-07T08:05:30.8316738Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8316920Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8316991Z return mod(**inputs) 2025-09-07T08:05:30.8317223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8317307Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8317532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8317590Z outputs = block( 2025-09-07T08:05:30.8317801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8317877Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8318111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8318175Z return func(*args, **kwargs) 2025-09-07T08:05:30.8318402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8318511Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8318735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8318817Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8319019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8319138Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8319141Z 2025-09-07T08:05:30.8319239Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8319421Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8319495Z return mod(**inputs) 2025-09-07T08:05:30.8319726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8319813Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8320040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8320099Z outputs = block( 2025-09-07T08:05:30.8320311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8320384Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8320614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8320706Z return func(*args, **kwargs) 2025-09-07T08:05:30.8320932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8321036Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8321290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8321376Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8321581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8321696Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8321699Z 2025-09-07T08:05:30.8321795Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8321978Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8322052Z return mod(**inputs) 2025-09-07T08:05:30.8322286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8322372Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8322599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8322660Z outputs = block( 2025-09-07T08:05:30.8322874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8322969Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8323203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8323269Z return func(*args, **kwargs) 2025-09-07T08:05:30.8323509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8323604Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8323830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:05:30.8323912Z hidden_states = self.act(hidden_states) 2025-09-07T08:05:30.8324113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:30.8324291Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:30.8324294Z 2025-09-07T08:05:30.8324391Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8324574Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8324641Z return mod(**inputs) 2025-09-07T08:05:30.8324878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8324962Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8325192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8325258Z outputs = block( 2025-09-07T08:05:30.8325464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8325537Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8325773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8325837Z return func(*args, **kwargs) 2025-09-07T08:05:30.8326072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8326165Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8326424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8326515Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8326716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8326867Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8326870Z 2025-09-07T08:05:30.8326967Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8327149Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8327218Z return mod(**inputs) 2025-09-07T08:05:30.8327446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8327531Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8327759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8327825Z outputs = block( 2025-09-07T08:05:30.8328027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8328100Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8328330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8328393Z return func(*args, **kwargs) 2025-09-07T08:05:30.8328625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8328717Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8328940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8329027Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8329230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8329345Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8329348Z 2025-09-07T08:05:30.8329443Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8329638Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8329699Z return mod(**inputs) 2025-09-07T08:05:30.8329927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8330013Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8330235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8330300Z outputs = block( 2025-09-07T08:05:30.8330504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8330576Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8330806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8330874Z return func(*args, **kwargs) 2025-09-07T08:05:30.8331102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8331183Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8331404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8331474Z return func(*args, **kwargs) 2025-09-07T08:05:30.8331701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8331907Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8332110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8332227Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8332259Z 2025-09-07T08:05:30.8332355Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8332538Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8332609Z return mod(**inputs) 2025-09-07T08:05:30.8332838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8332923Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8333148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8333208Z outputs = block( 2025-09-07T08:05:30.8333423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8333497Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8333728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8333795Z return func(*args, **kwargs) 2025-09-07T08:05:30.8334022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8334110Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8334329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8334403Z return func(*args, **kwargs) 2025-09-07T08:05:30.8334631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8334813Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8335018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8335127Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8335133Z 2025-09-07T08:05:30.8335220Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8335296Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8335380Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8335449Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8335548Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8335739Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8335802Z return mod(**inputs) 2025-09-07T08:05:30.8336043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8336120Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8336346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8336415Z outputs = block( 2025-09-07T08:05:30.8336619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8336699Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8336920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8336992Z return func(*args, **kwargs) 2025-09-07T08:05:30.8337217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8337298Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8337558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8337623Z return func(*args, **kwargs) 2025-09-07T08:05:30.8337858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8337979Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8338248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:05:30.8338377Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:30.8338379Z 2025-09-07T08:05:30.8338475Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8338666Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8338729Z return mod(**inputs) 2025-09-07T08:05:30.8338968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8339052Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8339277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8339350Z outputs = block( 2025-09-07T08:05:30.8339555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8339636Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8339857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8339921Z return func(*args, **kwargs) 2025-09-07T08:05:30.8340157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8340240Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8340471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8340534Z return func(*args, **kwargs) 2025-09-07T08:05:30.8340761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8340866Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8341135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:05:30.8341246Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:05:30.8341250Z 2025-09-07T08:05:30.8341345Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8341535Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8341595Z return mod(**inputs) 2025-09-07T08:05:30.8341832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8341916Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8342141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8342210Z outputs = block( 2025-09-07T08:05:30.8342416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8342489Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8342719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8342781Z return func(*args, **kwargs) 2025-09-07T08:05:30.8343017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8343122Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8343344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8343414Z return func(*args, **kwargs) 2025-09-07T08:05:30.8343639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8343752Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8343954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8344069Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8344072Z 2025-09-07T08:05:30.8344168Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8344351Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8344422Z return mod(**inputs) 2025-09-07T08:05:30.8344656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8344738Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8344964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8345024Z outputs = block( 2025-09-07T08:05:30.8345235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8345306Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8345536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8345599Z return func(*args, **kwargs) 2025-09-07T08:05:30.8345827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8345918Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8346138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8346209Z return func(*args, **kwargs) 2025-09-07T08:05:30.8346435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8346520Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8346722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8346830Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8346833Z 2025-09-07T08:05:30.8346938Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8347122Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8347191Z return mod(**inputs) 2025-09-07T08:05:30.8347425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8347504Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8347737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8347798Z outputs = block( 2025-09-07T08:05:30.8348012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8348086Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8348308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8348382Z return func(*args, **kwargs) 2025-09-07T08:05:30.8348611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8348744Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8348972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8349056Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8349300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8349408Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8349411Z 2025-09-07T08:05:30.8349516Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8349698Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8349768Z return mod(**inputs) 2025-09-07T08:05:30.8350000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8350079Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8350315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8350373Z outputs = block( 2025-09-07T08:05:30.8350584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8350661Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8350889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8350955Z return func(*args, **kwargs) 2025-09-07T08:05:30.8351183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8351287Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8351517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8351600Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8351801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8351909Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8351915Z 2025-09-07T08:05:30.8352021Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8352208Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8352278Z return mod(**inputs) 2025-09-07T08:05:30.8352509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8352585Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8352820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8352883Z outputs = block( 2025-09-07T08:05:30.8353094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8353165Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8353392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8353461Z return func(*args, **kwargs) 2025-09-07T08:05:30.8353686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8353789Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8354016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:05:30.8354100Z hidden_states = self.act(hidden_states) 2025-09-07T08:05:30.8354327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:30.8354496Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:30.8354505Z 2025-09-07T08:05:30.8354601Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8354821Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8354892Z return mod(**inputs) 2025-09-07T08:05:30.8355124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8355211Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8355438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8355496Z outputs = block( 2025-09-07T08:05:30.8355706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8355779Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8356012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8356075Z return func(*args, **kwargs) 2025-09-07T08:05:30.8356302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8356405Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8356627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8356715Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8356916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8357026Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8357036Z 2025-09-07T08:05:30.8357132Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8357314Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8357387Z return mod(**inputs) 2025-09-07T08:05:30.8357624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8357709Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8357936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8357995Z outputs = block( 2025-09-07T08:05:30.8358205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8358277Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8358510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8358574Z return func(*args, **kwargs) 2025-09-07T08:05:30.8358801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8358903Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8359128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8359217Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8359417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8359531Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8359534Z 2025-09-07T08:05:30.8359631Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8359877Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8359951Z return mod(**inputs) 2025-09-07T08:05:30.8360184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8360268Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8360524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8360583Z outputs = block( 2025-09-07T08:05:30.8360798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8360872Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8361102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8361166Z return func(*args, **kwargs) 2025-09-07T08:05:30.8361399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 442, in forward 2025-09-07T08:05:30.8361507Z hidden_states = residual + feed_forward_hidden_states 2025-09-07T08:05:30.8361510Z 2025-09-07T08:05:30.8361608Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8361807Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8361868Z return mod(**inputs) 2025-09-07T08:05:30.8362108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8362189Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8362418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8362489Z outputs = block( 2025-09-07T08:05:30.8362698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8362780Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8363008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8363073Z return func(*args, **kwargs) 2025-09-07T08:05:30.8363311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8363392Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8363623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8363685Z return func(*args, **kwargs) 2025-09-07T08:05:30.8363910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8364089Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8364293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8364406Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8364409Z 2025-09-07T08:05:30.8364508Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8364696Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8364758Z return mod(**inputs) 2025-09-07T08:05:30.8364989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8365074Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8365302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8365368Z outputs = block( 2025-09-07T08:05:30.8365598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8365674Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8365902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8366295Z return func(*args, **kwargs) 2025-09-07T08:05:30.8366530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8366615Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8366841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8366906Z return func(*args, **kwargs) 2025-09-07T08:05:30.8367132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:05:30.8367312Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:05:30.8367512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8367629Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8367634Z 2025-09-07T08:05:30.8367713Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8367787Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8367865Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8367936Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8368041Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8368222Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8368317Z return mod(**inputs) 2025-09-07T08:05:30.8368561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8368636Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8368870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8368927Z outputs = block( 2025-09-07T08:05:30.8369144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8369217Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8369440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8369512Z return func(*args, **kwargs) 2025-09-07T08:05:30.8369740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8369828Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8370056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8370121Z return func(*args, **kwargs) 2025-09-07T08:05:30.8370354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8370450Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8370727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:05:30.8370844Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:30.8370847Z 2025-09-07T08:05:30.8370951Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8371137Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8371200Z return mod(**inputs) 2025-09-07T08:05:30.8371495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8371573Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8371810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8371906Z outputs = block( 2025-09-07T08:05:30.8372109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8372190Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8372413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8372482Z return func(*args, **kwargs) 2025-09-07T08:05:30.8372708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8372790Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8373023Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8373089Z return func(*args, **kwargs) 2025-09-07T08:05:30.8373328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:05:30.8373424Z attn_output, attn_weights = attention_interface( 2025-09-07T08:05:30.8373691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:05:30.8373800Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:05:30.8373803Z 2025-09-07T08:05:30.8373896Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8374088Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8374151Z return mod(**inputs) 2025-09-07T08:05:30.8374397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8374475Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8374701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8374772Z outputs = block( 2025-09-07T08:05:30.8374974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8375056Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8375275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8375341Z return func(*args, **kwargs) 2025-09-07T08:05:30.8375575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8375657Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8375889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8375952Z return func(*args, **kwargs) 2025-09-07T08:05:30.8376186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8376266Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8376466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8376582Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8376585Z 2025-09-07T08:05:30.8376679Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8376866Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8376927Z return mod(**inputs) 2025-09-07T08:05:30.8377183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8377269Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8377496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8377590Z outputs = block( 2025-09-07T08:05:30.8377800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8377873Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8378106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8378169Z return func(*args, **kwargs) 2025-09-07T08:05:30.8378406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:05:30.8378488Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:05:30.8378719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8378786Z return func(*args, **kwargs) 2025-09-07T08:05:30.8379018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:05:30.8379105Z attn_output = self.c_proj(attn_output) 2025-09-07T08:05:30.8379310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8379424Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8379427Z 2025-09-07T08:05:30.8379523Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8379709Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8379777Z return mod(**inputs) 2025-09-07T08:05:30.8380015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8380097Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8380332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8380393Z outputs = block( 2025-09-07T08:05:30.8380610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8380684Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8380918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8380980Z return func(*args, **kwargs) 2025-09-07T08:05:30.8381216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8381314Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8381546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8381629Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8381833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8381950Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8381953Z 2025-09-07T08:05:30.8382052Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8382236Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8382303Z return mod(**inputs) 2025-09-07T08:05:30.8382537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8382624Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8382886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8382952Z outputs = block( 2025-09-07T08:05:30.8383155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8383252Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8383481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8383545Z return func(*args, **kwargs) 2025-09-07T08:05:30.8383776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8383870Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8384095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:05:30.8384181Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:05:30.8384381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8384495Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8384501Z 2025-09-07T08:05:30.8384596Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8384780Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8384849Z return mod(**inputs) 2025-09-07T08:05:30.8385077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8385163Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8385390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8385457Z outputs = block( 2025-09-07T08:05:30.8385663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8385734Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8385964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8386030Z return func(*args, **kwargs) 2025-09-07T08:05:30.8386265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8386359Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8386589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:05:30.8386670Z hidden_states = self.act(hidden_states) 2025-09-07T08:05:30.8386870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:30.8387047Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:30.8387050Z 2025-09-07T08:05:30.8387146Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8387336Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8387401Z return mod(**inputs) 2025-09-07T08:05:30.8387632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8387720Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8387948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8388015Z outputs = block( 2025-09-07T08:05:30.8388217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8388316Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8388551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8388614Z return func(*args, **kwargs) 2025-09-07T08:05:30.8388873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8388966Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8389192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8389283Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8389484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8389600Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8389603Z 2025-09-07T08:05:30.8389702Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:30.8389893Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:30.8389955Z return mod(**inputs) 2025-09-07T08:05:30.8390185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:05:30.8390273Z transformer_outputs = self.transformer( 2025-09-07T08:05:30.8390497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:05:30.8390562Z outputs = block( 2025-09-07T08:05:30.8390767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:05:30.8390843Z return super().__call__(*args, **kwargs) 2025-09-07T08:05:30.8391079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:05:30.8391145Z return func(*args, **kwargs) 2025-09-07T08:05:30.8391377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:05:30.8391472Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:05:30.8391706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:05:30.8391788Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:05:30.8391990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:05:30.8392106Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:05:30.8392109Z 2025-09-07T08:05:30.8392185Z cudagraph partition due to non gpu ops 2025-09-07T08:05:30.8392271Z cudagraph partition due to non gpu ops 2025-09-07T08:05:46.8361642Z pass 2025-09-07T08:05:46.8362087Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:48.8237930Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:05:48.8239502Z import pynvml # type: ignore[import] 2025-09-07T08:05:50.5400233Z 2025-09-07T08:06:01.2485599Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:06:01.2485912Z loading model: 0it [00:10, ?it/s] 2025-09-07T08:06:01.2486165Z cpu eval hf_GPT2_large 2025-09-07T08:06:01.6224800Z pass_due_to_skip 2025-09-07T08:06:01.6225320Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:06:03.1224883Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:06:03.1226287Z import pynvml # type: ignore[import] 2025-09-07T08:06:04.8448243Z 2025-09-07T08:06:07.9155900Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:06:07.9156244Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:06:07.9156513Z cpu eval hf_Longformer 2025-09-07T08:06:08.9203458Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:06:10.0574331Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:06:11.1966723Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:06:11.4363896Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T08:06:11.4364945Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T08:06:11.4365822Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] or: 2025-09-07T08:06:11.4366614Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T08:06:11.4367437Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] to include these operations in the captured graph. 2025-09-07T08:06:11.4368172Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:06:11.4368957Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break: from user code at: 2025-09-07T08:06:11.4370087Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:11.4371108Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] return mod(**inputs) 2025-09-07T08:06:11.4372295Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1703, in forward 2025-09-07T08:06:11.4373463Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] outputs = self.longformer( 2025-09-07T08:06:11.4374653Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1600, in forward 2025-09-07T08:06:11.4375852Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] encoder_outputs = self.encoder( 2025-09-07T08:06:11.4377086Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1244, in forward 2025-09-07T08:06:11.4378399Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] is_global_attn = is_index_global_attn.flatten().any().item() 2025-09-07T08:06:11.4379194Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:06:11.4379777Z W0907 08:06:11.435087 44289 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:06:21.0119113Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:21.0119674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1596, in forward 2025-09-07T08:06:21.0120097Z embedding_output = self.embeddings( 2025-09-07T08:06:21.0120793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 410, in forward 2025-09-07T08:06:21.0121331Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx).to(input_ids.device) 2025-09-07T08:06:21.0122011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 381, in create_position_ids_from_input_ids 2025-09-07T08:06:21.0122474Z mask = input_ids.ne(padding_idx).int() 2025-09-07T08:06:21.0122611Z 2025-09-07T08:06:21.0122724Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:21.0123145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1592, in forward 2025-09-07T08:06:21.0123645Z extended_attention_mask: torch.Tensor = self.get_extended_attention_mask(attention_mask, input_shape)[ 2025-09-07T08:06:21.0124150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_utils.py", line 1763, in get_extended_attention_mask 2025-09-07T08:06:21.0124634Z extended_attention_mask = (1.0 - extended_attention_mask) * torch.finfo(dtype).min 2025-09-07T08:06:21.0124847Z 2025-09-07T08:06:21.0124947Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:21.0125365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1596, in forward 2025-09-07T08:06:21.0125763Z embedding_output = self.embeddings( 2025-09-07T08:06:21.0126146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 410, in forward 2025-09-07T08:06:21.0126642Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx).to(input_ids.device) 2025-09-07T08:06:21.0127196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 382, in create_position_ids_from_input_ids 2025-09-07T08:06:21.0127699Z incremental_indices = torch.cumsum(mask, dim=1).type_as(mask) * mask 2025-09-07T08:06:21.0127878Z 2025-09-07T08:06:21.0127985Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:21.0128393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1596, in forward 2025-09-07T08:06:21.0128788Z embedding_output = self.embeddings( 2025-09-07T08:06:21.0129178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 410, in forward 2025-09-07T08:06:21.0129668Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx).to(input_ids.device) 2025-09-07T08:06:21.0130212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 382, in create_position_ids_from_input_ids 2025-09-07T08:06:21.0130704Z incremental_indices = torch.cumsum(mask, dim=1).type_as(mask) * mask 2025-09-07T08:06:21.0130891Z 2025-09-07T08:06:59.5364190Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5364826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5365334Z layer_outputs = layer_module( 2025-09-07T08:06:59.5365691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5366055Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5366476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5366884Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5367557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5367973Z self_outputs = self.self( 2025-09-07T08:06:59.5368444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5368856Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5369099Z 2025-09-07T08:06:59.5369217Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5369695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5370159Z layer_outputs = layer_module( 2025-09-07T08:06:59.5370496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5370843Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5371236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5371639Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5372030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5372420Z self_outputs = self.self( 2025-09-07T08:06:59.5372795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T08:06:59.5373214Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5373695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 787, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5374344Z key = key.transpose(1, 2).reshape(batch_size * num_heads, seq_len, head_dim) 2025-09-07T08:06:59.5374538Z 2025-09-07T08:06:59.5374639Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5375120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5375574Z layer_outputs = layer_module( 2025-09-07T08:06:59.5375892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5376242Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5376628Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5377019Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5377397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5377781Z self_outputs = self.self( 2025-09-07T08:06:59.5378156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5378556Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5378688Z 2025-09-07T08:06:59.5378792Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5379262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5379718Z layer_outputs = layer_module( 2025-09-07T08:06:59.5380050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5380398Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5380790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5381172Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5381621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5382012Z self_outputs = self.self( 2025-09-07T08:06:59.5382388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5382854Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5383324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 789, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5383857Z query = self._chunk(query, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:06:59.5384336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:06:59.5384734Z hidden_states = hidden_states.view( 2025-09-07T08:06:59.5384857Z 2025-09-07T08:06:59.5384967Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5385449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5385911Z layer_outputs = layer_module( 2025-09-07T08:06:59.5386252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5386593Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5386989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5387373Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5387765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5388157Z self_outputs = self.self( 2025-09-07T08:06:59.5388539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:06:59.5388938Z key_vectors = self.key(hidden_states) 2025-09-07T08:06:59.5389067Z 2025-09-07T08:06:59.5389164Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5389657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5390109Z layer_outputs = layer_module( 2025-09-07T08:06:59.5390439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5390776Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5391170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5391569Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5391964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5392353Z self_outputs = self.self( 2025-09-07T08:06:59.5392722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5393144Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5393611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5394129Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:06:59.5394592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:06:59.5394983Z hidden_states = hidden_states.view( 2025-09-07T08:06:59.5395140Z 2025-09-07T08:06:59.5395242Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5395727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5396219Z layer_outputs = layer_module( 2025-09-07T08:06:59.5396555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5396890Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5397287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5397686Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5398077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5398465Z self_outputs = self.self( 2025-09-07T08:06:59.5398837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5399251Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5399725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5400279Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5400509Z 2025-09-07T08:06:59.5400596Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5400804Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5401005Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5401198Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5401417Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5401896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5402357Z layer_outputs = layer_module( 2025-09-07T08:06:59.5402691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5403039Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5403433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5403822Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5404213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5404601Z self_outputs = self.self( 2025-09-07T08:06:59.5404985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T08:06:59.5405422Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5405924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5406441Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T08:06:59.5406926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T08:06:59.5407423Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T08:06:59.5407619Z 2025-09-07T08:06:59.5407695Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5407924Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5408432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5408883Z layer_outputs = layer_module( 2025-09-07T08:06:59.5409212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5409601Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5409992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5410370Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5410761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5411150Z self_outputs = self.self( 2025-09-07T08:06:59.5411523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:06:59.5411905Z attn_scores += diagonal_mask 2025-09-07T08:06:59.5412021Z 2025-09-07T08:06:59.5412118Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5412596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5413048Z layer_outputs = layer_module( 2025-09-07T08:06:59.5413376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5413710Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5414092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5414477Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5414860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5415243Z self_outputs = self.self( 2025-09-07T08:06:59.5415614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:06:59.5415995Z attn_probs = nn.functional.softmax( 2025-09-07T08:06:59.5416129Z 2025-09-07T08:06:59.5416226Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5416701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5417155Z layer_outputs = layer_module( 2025-09-07T08:06:59.5417479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5417810Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5418204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5418592Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5418980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5419359Z self_outputs = self.self( 2025-09-07T08:06:59.5419735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:06:59.5420136Z value_vectors = self.value(hidden_states) 2025-09-07T08:06:59.5420266Z 2025-09-07T08:06:59.5420372Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5420847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5421287Z layer_outputs = layer_module( 2025-09-07T08:06:59.5421647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5421991Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5422388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5422808Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5423190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5423576Z self_outputs = self.self( 2025-09-07T08:06:59.5423952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5424378Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5424879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5425423Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:06:59.5425827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5426161Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5426308Z 2025-09-07T08:06:59.5426412Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5426899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5427346Z layer_outputs = layer_module( 2025-09-07T08:06:59.5427679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5428024Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5428419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5428811Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5429195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5429584Z self_outputs = self.self( 2025-09-07T08:06:59.5429960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5430384Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5430867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5431382Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:06:59.5431862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:06:59.5432308Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:06:59.5432628Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5432957Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5433108Z 2025-09-07T08:06:59.5433205Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5433691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5434147Z layer_outputs = layer_module( 2025-09-07T08:06:59.5434477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5434811Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5435234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5435625Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5436012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5436426Z self_outputs = self.self( 2025-09-07T08:06:59.5436792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5437220Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5437707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5438234Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5438424Z 2025-09-07T08:06:59.5438535Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5439007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5439456Z layer_outputs = layer_module( 2025-09-07T08:06:59.5439789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5440127Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5440518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5440899Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5441285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5441667Z self_outputs = self.self( 2025-09-07T08:06:59.5442040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5442459Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5442940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5443459Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5443657Z 2025-09-07T08:06:59.5443755Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5444233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5444684Z layer_outputs = layer_module( 2025-09-07T08:06:59.5445005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5445345Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5445738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5446127Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5446511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5446894Z self_outputs = self.self( 2025-09-07T08:06:59.5447269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T08:06:59.5447754Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T08:06:59.5447975Z 2025-09-07T08:06:59.5448080Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5450273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5450741Z layer_outputs = layer_module( 2025-09-07T08:06:59.5451066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5451460Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5451852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5452229Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5452612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:06:59.5453033Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:59.5453456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:06:59.5453851Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:59.5453980Z 2025-09-07T08:06:59.5454056Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5454258Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5454458Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5454649Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5454860Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5455339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5455786Z layer_outputs = layer_module( 2025-09-07T08:06:59.5456110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5456445Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5456835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5457227Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5457613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5458000Z self_outputs = self.self( 2025-09-07T08:06:59.5458371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5458764Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5458904Z 2025-09-07T08:06:59.5459002Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5459478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5459936Z layer_outputs = layer_module( 2025-09-07T08:06:59.5460266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5460596Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5460986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5461378Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5461765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5462145Z self_outputs = self.self( 2025-09-07T08:06:59.5462520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5462909Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5463037Z 2025-09-07T08:06:59.5463191Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5463673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5464115Z layer_outputs = layer_module( 2025-09-07T08:06:59.5464471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5464808Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5465197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5465588Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5465970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5466352Z self_outputs = self.self( 2025-09-07T08:06:59.5466727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5467143Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5467614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5468157Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5468438Z 2025-09-07T08:06:59.5468537Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5469021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5469488Z layer_outputs = layer_module( 2025-09-07T08:06:59.5469830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5470178Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5470581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5470982Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5471387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5471784Z self_outputs = self.self( 2025-09-07T08:06:59.5472162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:06:59.5472566Z key_vectors = self.key(hidden_states) 2025-09-07T08:06:59.5472702Z 2025-09-07T08:06:59.5472803Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5473295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5473759Z layer_outputs = layer_module( 2025-09-07T08:06:59.5474089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5474436Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5474844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5475246Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5475641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5476036Z self_outputs = self.self( 2025-09-07T08:06:59.5476419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5476900Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5477379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5477903Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:06:59.5478430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:06:59.5478828Z hidden_states = hidden_states.view( 2025-09-07T08:06:59.5478953Z 2025-09-07T08:06:59.5479075Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5479580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5480042Z layer_outputs = layer_module( 2025-09-07T08:06:59.5480388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5480738Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5481143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5481547Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5481937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5482336Z self_outputs = self.self( 2025-09-07T08:06:59.5482713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5483137Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5483615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5484176Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5484415Z 2025-09-07T08:06:59.5484493Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5484705Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5484944Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5485430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5485892Z layer_outputs = layer_module( 2025-09-07T08:06:59.5486225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5486577Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5486990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5487379Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5487768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5488156Z self_outputs = self.self( 2025-09-07T08:06:59.5488533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:06:59.5488919Z attn_scores += diagonal_mask 2025-09-07T08:06:59.5489034Z 2025-09-07T08:06:59.5489132Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5489617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5490069Z layer_outputs = layer_module( 2025-09-07T08:06:59.5490425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5490763Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5491146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5491536Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5491956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5492336Z self_outputs = self.self( 2025-09-07T08:06:59.5492698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:06:59.5493087Z attn_probs = nn.functional.softmax( 2025-09-07T08:06:59.5493220Z 2025-09-07T08:06:59.5493318Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5493806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5494249Z layer_outputs = layer_module( 2025-09-07T08:06:59.5494566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5494901Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5495296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5495682Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5496065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5496437Z self_outputs = self.self( 2025-09-07T08:06:59.5496811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:06:59.5497205Z value_vectors = self.value(hidden_states) 2025-09-07T08:06:59.5497334Z 2025-09-07T08:06:59.5497439Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5497914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5498365Z layer_outputs = layer_module( 2025-09-07T08:06:59.5498690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5499026Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5499411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5499797Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5500177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5500563Z self_outputs = self.self( 2025-09-07T08:06:59.5500931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5501356Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5501839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5502375Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:06:59.5502772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5503103Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5503243Z 2025-09-07T08:06:59.5503346Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5503847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5504304Z layer_outputs = layer_module( 2025-09-07T08:06:59.5504639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5505035Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5505426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5505811Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5506198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5506583Z self_outputs = self.self( 2025-09-07T08:06:59.5506957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5507389Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5507871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5508386Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:06:59.5508868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:06:59.5509312Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:06:59.5509632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5509954Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5510103Z 2025-09-07T08:06:59.5510201Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5510686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5511141Z layer_outputs = layer_module( 2025-09-07T08:06:59.5511470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5511809Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5512200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5512589Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5512979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5513360Z self_outputs = self.self( 2025-09-07T08:06:59.5513733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5514160Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5514649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5515187Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5515378Z 2025-09-07T08:06:59.5515485Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5515959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5516411Z layer_outputs = layer_module( 2025-09-07T08:06:59.5516739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5517080Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5517511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5517894Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5518284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5518699Z self_outputs = self.self( 2025-09-07T08:06:59.5519077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5519495Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5519983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5520502Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5520705Z 2025-09-07T08:06:59.5520803Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5521284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5521744Z layer_outputs = layer_module( 2025-09-07T08:06:59.5522061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5522400Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5522795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5523188Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5523568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5523956Z self_outputs = self.self( 2025-09-07T08:06:59.5524330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T08:06:59.5524823Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T08:06:59.5525048Z 2025-09-07T08:06:59.5525153Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5525626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5526078Z layer_outputs = layer_module( 2025-09-07T08:06:59.5526407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5526748Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5527145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5527526Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5527911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:06:59.5528336Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:59.5528755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:06:59.5529157Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:59.5529287Z 2025-09-07T08:06:59.5529365Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5529571Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5529771Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5529967Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5530180Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5530689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5531140Z layer_outputs = layer_module( 2025-09-07T08:06:59.5531466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5531842Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5532236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5532632Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5533024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5533418Z self_outputs = self.self( 2025-09-07T08:06:59.5533792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5534199Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5534336Z 2025-09-07T08:06:59.5534437Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5534930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5535398Z layer_outputs = layer_module( 2025-09-07T08:06:59.5535726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5536075Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5536475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5536871Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5537269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5537656Z self_outputs = self.self( 2025-09-07T08:06:59.5538041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5538449Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5538581Z 2025-09-07T08:06:59.5538688Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5539175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5539623Z layer_outputs = layer_module( 2025-09-07T08:06:59.5539958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5540303Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5540703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5541102Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5541486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5541883Z self_outputs = self.self( 2025-09-07T08:06:59.5542258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5542680Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5543149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5543705Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5543950Z 2025-09-07T08:06:59.5544146Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5544647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5545138Z layer_outputs = layer_module( 2025-09-07T08:06:59.5545473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5545812Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5546215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5546618Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5547021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5547412Z self_outputs = self.self( 2025-09-07T08:06:59.5547797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:06:59.5548195Z key_vectors = self.key(hidden_states) 2025-09-07T08:06:59.5548324Z 2025-09-07T08:06:59.5548434Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5548926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5549380Z layer_outputs = layer_module( 2025-09-07T08:06:59.5549712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5550059Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5550471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5550873Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5551259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5551651Z self_outputs = self.self( 2025-09-07T08:06:59.5552033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5552461Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5552927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5553444Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:06:59.5553916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:06:59.5554312Z hidden_states = hidden_states.view( 2025-09-07T08:06:59.5554437Z 2025-09-07T08:06:59.5554547Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5555037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5555489Z layer_outputs = layer_module( 2025-09-07T08:06:59.5555822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5556166Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5556562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5556959Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5557343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5557767Z self_outputs = self.self( 2025-09-07T08:06:59.5558148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5558562Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5559077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5559647Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5559890Z 2025-09-07T08:06:59.5559970Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5560182Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5560416Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5560912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5561386Z layer_outputs = layer_module( 2025-09-07T08:06:59.5561726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5562080Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5562491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5562891Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5563295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5563696Z self_outputs = self.self( 2025-09-07T08:06:59.5564085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:06:59.5564488Z attn_scores += diagonal_mask 2025-09-07T08:06:59.5564612Z 2025-09-07T08:06:59.5564720Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5565221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5565696Z layer_outputs = layer_module( 2025-09-07T08:06:59.5566040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5566387Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5566797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5567205Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5567613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5568019Z self_outputs = self.self( 2025-09-07T08:06:59.5568439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:06:59.5568849Z attn_probs = nn.functional.softmax( 2025-09-07T08:06:59.5568985Z 2025-09-07T08:06:59.5569090Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5569586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5570050Z layer_outputs = layer_module( 2025-09-07T08:06:59.5570383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5570737Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5571145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5571600Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5571998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5572385Z self_outputs = self.self( 2025-09-07T08:06:59.5572816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:06:59.5573223Z value_vectors = self.value(hidden_states) 2025-09-07T08:06:59.5573356Z 2025-09-07T08:06:59.5573476Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5573959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5574407Z layer_outputs = layer_module( 2025-09-07T08:06:59.5574747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5575092Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5575490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5575871Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5576262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5576644Z self_outputs = self.self( 2025-09-07T08:06:59.5577014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5577438Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5577921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5578464Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:06:59.5578861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5579192Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5579340Z 2025-09-07T08:06:59.5579444Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5579917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5580364Z layer_outputs = layer_module( 2025-09-07T08:06:59.5580688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5581030Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5581427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5581808Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5582192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5582576Z self_outputs = self.self( 2025-09-07T08:06:59.5582946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5583374Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5583855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5584362Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:06:59.5584863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:06:59.5585306Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:06:59.5585628Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5585952Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5586134Z 2025-09-07T08:06:59.5586235Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5586709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5587160Z layer_outputs = layer_module( 2025-09-07T08:06:59.5587484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5587815Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5588205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5588591Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5588982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5589354Z self_outputs = self.self( 2025-09-07T08:06:59.5589729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5590149Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5590632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5591148Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5591337Z 2025-09-07T08:06:59.5591451Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5591920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5592370Z layer_outputs = layer_module( 2025-09-07T08:06:59.5592701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5593041Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5593419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5593803Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5594189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5594574Z self_outputs = self.self( 2025-09-07T08:06:59.5594942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5595358Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5595840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5596360Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5596546Z 2025-09-07T08:06:59.5596652Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5597126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5597565Z layer_outputs = layer_module( 2025-09-07T08:06:59.5597894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5598253Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5598645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5599033Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5599440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5599826Z self_outputs = self.self( 2025-09-07T08:06:59.5600201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T08:06:59.5600690Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T08:06:59.5600908Z 2025-09-07T08:06:59.5601015Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5601484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5601964Z layer_outputs = layer_module( 2025-09-07T08:06:59.5602291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5602637Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5603026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5603410Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5603800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:06:59.5604222Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:59.5604646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:06:59.5605045Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:59.5605176Z 2025-09-07T08:06:59.5605251Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5605459Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5605659Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5605854Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5606066Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5606541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5606995Z layer_outputs = layer_module( 2025-09-07T08:06:59.5607321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5607662Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5608053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5608445Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5608829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5609218Z self_outputs = self.self( 2025-09-07T08:06:59.5609581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5609977Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5610114Z 2025-09-07T08:06:59.5610211Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5610696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5611142Z layer_outputs = layer_module( 2025-09-07T08:06:59.5611484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5611827Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5612219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5612650Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5613032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5613406Z self_outputs = self.self( 2025-09-07T08:06:59.5613777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5614172Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5614298Z 2025-09-07T08:06:59.5614403Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5614883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5615322Z layer_outputs = layer_module( 2025-09-07T08:06:59.5615645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5615984Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5616373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5616757Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5617135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5617515Z self_outputs = self.self( 2025-09-07T08:06:59.5617886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5618301Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5618752Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5619302Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5619536Z 2025-09-07T08:06:59.5619635Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5620111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5620560Z layer_outputs = layer_module( 2025-09-07T08:06:59.5620879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5621223Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5621623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5622008Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5622394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5622766Z self_outputs = self.self( 2025-09-07T08:06:59.5623134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:06:59.5623523Z key_vectors = self.key(hidden_states) 2025-09-07T08:06:59.5623648Z 2025-09-07T08:06:59.5623756Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5624254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5624705Z layer_outputs = layer_module( 2025-09-07T08:06:59.5625038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5625377Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5625796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5626185Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5626562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5626949Z self_outputs = self.self( 2025-09-07T08:06:59.5627314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5627722Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5628186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5628690Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:06:59.5629148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:06:59.5629531Z hidden_states = hidden_states.view( 2025-09-07T08:06:59.5629653Z 2025-09-07T08:06:59.5629759Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5630236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5630675Z layer_outputs = layer_module( 2025-09-07T08:06:59.5631006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5631345Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5631733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5632110Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5632500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5632880Z self_outputs = self.self( 2025-09-07T08:06:59.5633250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5633660Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5634112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5634656Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5634890Z 2025-09-07T08:06:59.5634967Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5635171Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5635396Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5635872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5636324Z layer_outputs = layer_module( 2025-09-07T08:06:59.5636651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5636991Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5637378Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5637779Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5638170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5638553Z self_outputs = self.self( 2025-09-07T08:06:59.5638929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:06:59.5639343Z attn_scores += diagonal_mask 2025-09-07T08:06:59.5639471Z 2025-09-07T08:06:59.5639570Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5640051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5640504Z layer_outputs = layer_module( 2025-09-07T08:06:59.5640831Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5641167Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5641564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5641955Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5642355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5642745Z self_outputs = self.self( 2025-09-07T08:06:59.5643113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:06:59.5643507Z attn_probs = nn.functional.softmax( 2025-09-07T08:06:59.5643640Z 2025-09-07T08:06:59.5643738Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5644220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5644675Z layer_outputs = layer_module( 2025-09-07T08:06:59.5644996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5645338Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5645735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5646133Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5646524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5646908Z self_outputs = self.self( 2025-09-07T08:06:59.5647291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:06:59.5647689Z value_vectors = self.value(hidden_states) 2025-09-07T08:06:59.5647823Z 2025-09-07T08:06:59.5647931Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5648406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5648865Z layer_outputs = layer_module( 2025-09-07T08:06:59.5649192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5649534Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5649931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5650319Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5650710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5651134Z self_outputs = self.self( 2025-09-07T08:06:59.5651511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5651941Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5652453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5652995Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:06:59.5653391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5653728Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5653871Z 2025-09-07T08:06:59.5653979Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5654448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5654892Z layer_outputs = layer_module( 2025-09-07T08:06:59.5655221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5655563Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5655957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5656340Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5656731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5657115Z self_outputs = self.self( 2025-09-07T08:06:59.5657485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5657910Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5658395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5658903Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:06:59.5659378Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:06:59.5659815Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:06:59.5660134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5660454Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5660606Z 2025-09-07T08:06:59.5660706Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5661187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5661640Z layer_outputs = layer_module( 2025-09-07T08:06:59.5661961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5662303Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5662699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5663090Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5663474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5663850Z self_outputs = self.self( 2025-09-07T08:06:59.5664260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5664686Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5665180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5665733Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5665921Z 2025-09-07T08:06:59.5666018Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5666499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5666951Z layer_outputs = layer_module( 2025-09-07T08:06:59.5667274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5667612Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5668004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5668444Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5668844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5669249Z self_outputs = self.self( 2025-09-07T08:06:59.5669623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5670045Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5670546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5671083Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5671279Z 2025-09-07T08:06:59.5671391Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5671889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5672349Z layer_outputs = layer_module( 2025-09-07T08:06:59.5672690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5673040Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5673440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5673843Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5674234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5674636Z self_outputs = self.self( 2025-09-07T08:06:59.5675020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T08:06:59.5675527Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T08:06:59.5675757Z 2025-09-07T08:06:59.5675864Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5676350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5676810Z layer_outputs = layer_module( 2025-09-07T08:06:59.5677145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5677494Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5677948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5678347Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5678749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:06:59.5679186Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:59.5679658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:06:59.5680058Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:59.5680199Z 2025-09-07T08:06:59.5680277Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5680485Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5680693Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5680901Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5681118Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5681616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5682081Z layer_outputs = layer_module( 2025-09-07T08:06:59.5682405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5682740Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5683131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5683519Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5683910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5684291Z self_outputs = self.self( 2025-09-07T08:06:59.5684657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5685056Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5685191Z 2025-09-07T08:06:59.5685287Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5685766Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5686215Z layer_outputs = layer_module( 2025-09-07T08:06:59.5686529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5686864Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5687248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5687638Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5688026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5688397Z self_outputs = self.self( 2025-09-07T08:06:59.5688765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5689160Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5689288Z 2025-09-07T08:06:59.5689393Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5689871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5690313Z layer_outputs = layer_module( 2025-09-07T08:06:59.5690635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5690972Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5691388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5691773Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5692162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5692584Z self_outputs = self.self( 2025-09-07T08:06:59.5692959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5693375Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5693832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5694387Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5694621Z 2025-09-07T08:06:59.5694722Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5695208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5695661Z layer_outputs = layer_module( 2025-09-07T08:06:59.5695990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5696331Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5696724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5697117Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5697509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5697884Z self_outputs = self.self( 2025-09-07T08:06:59.5698264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:06:59.5698652Z key_vectors = self.key(hidden_states) 2025-09-07T08:06:59.5698776Z 2025-09-07T08:06:59.5698882Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5699372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5699813Z layer_outputs = layer_module( 2025-09-07T08:06:59.5700144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5700488Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5700886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5701279Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5701663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5702044Z self_outputs = self.self( 2025-09-07T08:06:59.5702412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5702825Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5703281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5703798Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:06:59.5704258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:06:59.5704677Z hidden_states = hidden_states.view( 2025-09-07T08:06:59.5704804Z 2025-09-07T08:06:59.5705148Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5705621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5706097Z layer_outputs = layer_module( 2025-09-07T08:06:59.5706424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5706763Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5707155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5707534Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5707919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5708308Z self_outputs = self.self( 2025-09-07T08:06:59.5708679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5709090Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5709548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5710093Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5710326Z 2025-09-07T08:06:59.5710404Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5710606Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5710832Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5711301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5711751Z layer_outputs = layer_module( 2025-09-07T08:06:59.5712073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5712408Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5712797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5713180Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5713564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5713949Z self_outputs = self.self( 2025-09-07T08:06:59.5714318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:06:59.5714695Z attn_scores += diagonal_mask 2025-09-07T08:06:59.5714820Z 2025-09-07T08:06:59.5714924Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5715396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5715848Z layer_outputs = layer_module( 2025-09-07T08:06:59.5716175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5716509Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5716904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5717291Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5717673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5718056Z self_outputs = self.self( 2025-09-07T08:06:59.5718455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:06:59.5718845Z attn_probs = nn.functional.softmax( 2025-09-07T08:06:59.5718976Z 2025-09-07T08:06:59.5719074Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5719599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5720049Z layer_outputs = layer_module( 2025-09-07T08:06:59.5720368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5720705Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5721095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5721490Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5721876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5722254Z self_outputs = self.self( 2025-09-07T08:06:59.5722625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:06:59.5723024Z value_vectors = self.value(hidden_states) 2025-09-07T08:06:59.5723153Z 2025-09-07T08:06:59.5723255Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5723724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5724173Z layer_outputs = layer_module( 2025-09-07T08:06:59.5724501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5724846Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5725239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5725620Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5726013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5726398Z self_outputs = self.self( 2025-09-07T08:06:59.5726772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5727202Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5727684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5728233Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:06:59.5728632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5728959Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5729105Z 2025-09-07T08:06:59.5729209Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5729676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5730125Z layer_outputs = layer_module( 2025-09-07T08:06:59.5730454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5730793Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5731210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5731593Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5731984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5732370Z self_outputs = self.self( 2025-09-07T08:06:59.5732774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5733203Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5733687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5734201Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:06:59.5734682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:06:59.5735124Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:06:59.5735442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5735766Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5735918Z 2025-09-07T08:06:59.5736016Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5736502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5736955Z layer_outputs = layer_module( 2025-09-07T08:06:59.5737273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5737614Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5738011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5738402Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5738785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5739168Z self_outputs = self.self( 2025-09-07T08:06:59.5739543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5739970Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5740455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5740977Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5741168Z 2025-09-07T08:06:59.5741266Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5741749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5742205Z layer_outputs = layer_module( 2025-09-07T08:06:59.5742532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5742881Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5743266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5743661Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5744049Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5744438Z self_outputs = self.self( 2025-09-07T08:06:59.5744844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5745267Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5745762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5746320Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5746508Z 2025-09-07T08:06:59.5746614Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5747093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5747537Z layer_outputs = layer_module( 2025-09-07T08:06:59.5747862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5748204Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5748597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5748983Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5749367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5749748Z self_outputs = self.self( 2025-09-07T08:06:59.5750122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T08:06:59.5750613Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T08:06:59.5750831Z 2025-09-07T08:06:59.5750937Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5751415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5751864Z layer_outputs = layer_module( 2025-09-07T08:06:59.5752193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5752534Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5752925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5753308Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5753700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:06:59.5754120Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:59.5754544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:06:59.5754943Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:59.5755083Z 2025-09-07T08:06:59.5755159Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5755360Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5755557Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5755754Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5755967Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5756451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5756906Z layer_outputs = layer_module( 2025-09-07T08:06:59.5757234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5757568Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5757990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5758381Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5758768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5759190Z self_outputs = self.self( 2025-09-07T08:06:59.5759557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5759955Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5760098Z 2025-09-07T08:06:59.5760195Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5760666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5761117Z layer_outputs = layer_module( 2025-09-07T08:06:59.5761441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5761789Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5762183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5762587Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5763001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5763388Z self_outputs = self.self( 2025-09-07T08:06:59.5763769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5764175Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5764307Z 2025-09-07T08:06:59.5764412Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5764903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5765362Z layer_outputs = layer_module( 2025-09-07T08:06:59.5765696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5766043Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5766313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5766392Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5766661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5766737Z self_outputs = self.self( 2025-09-07T08:06:59.5767011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5767111Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5767442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5767620Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5767623Z 2025-09-07T08:06:59.5767730Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5768065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5768142Z layer_outputs = layer_module( 2025-09-07T08:06:59.5768381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5768507Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5768782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5768854Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5769127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5769241Z self_outputs = self.self( 2025-09-07T08:06:59.5769501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:06:59.5769584Z key_vectors = self.key(hidden_states) 2025-09-07T08:06:59.5769588Z 2025-09-07T08:06:59.5769686Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5770030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5770101Z layer_outputs = layer_module( 2025-09-07T08:06:59.5770317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5770390Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5770655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5770736Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5770996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5771070Z self_outputs = self.self( 2025-09-07T08:06:59.5771329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5771426Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5771749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5771893Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:06:59.5772161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:06:59.5772234Z hidden_states = hidden_states.view( 2025-09-07T08:06:59.5772237Z 2025-09-07T08:06:59.5772341Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5772665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5772733Z layer_outputs = layer_module( 2025-09-07T08:06:59.5772949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5773024Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5773297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5773367Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5773636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5773701Z self_outputs = self.self( 2025-09-07T08:06:59.5773959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5774062Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5774373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5774588Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5774591Z 2025-09-07T08:06:59.5774668Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5774739Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5774848Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5775211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5775286Z layer_outputs = layer_module( 2025-09-07T08:06:59.5775493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5775576Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5775837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5775907Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5776173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5776237Z self_outputs = self.self( 2025-09-07T08:06:59.5776507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:06:59.5776577Z attn_scores += diagonal_mask 2025-09-07T08:06:59.5776580Z 2025-09-07T08:06:59.5776677Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5777007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5777075Z layer_outputs = layer_module( 2025-09-07T08:06:59.5777287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5777361Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5777632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5777703Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5777968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5778044Z self_outputs = self.self( 2025-09-07T08:06:59.5778303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:06:59.5778381Z attn_probs = nn.functional.softmax( 2025-09-07T08:06:59.5778384Z 2025-09-07T08:06:59.5778481Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5778810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5778884Z layer_outputs = layer_module( 2025-09-07T08:06:59.5779091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5779171Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5779434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5779509Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5779774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5779838Z self_outputs = self.self( 2025-09-07T08:06:59.5780102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:06:59.5780179Z value_vectors = self.value(hidden_states) 2025-09-07T08:06:59.5780182Z 2025-09-07T08:06:59.5780314Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5780637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5780740Z layer_outputs = layer_module( 2025-09-07T08:06:59.5780946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5781019Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5781288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5781358Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5781630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5781696Z self_outputs = self.self( 2025-09-07T08:06:59.5781964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5782085Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5782414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5782590Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:06:59.5782773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5782875Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5782878Z 2025-09-07T08:06:59.5782976Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5783310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5783386Z layer_outputs = layer_module( 2025-09-07T08:06:59.5783590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5783674Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5783934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5784004Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5784271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5784337Z self_outputs = self.self( 2025-09-07T08:06:59.5784601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5784713Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5785050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5785181Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:06:59.5785480Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:06:59.5785575Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:06:59.5785752Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5785853Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5785855Z 2025-09-07T08:06:59.5785952Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5786311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5786379Z layer_outputs = layer_module( 2025-09-07T08:06:59.5786583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5786706Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5786964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5787042Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5787299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5787363Z self_outputs = self.self( 2025-09-07T08:06:59.5787625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5787734Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5788067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5788212Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5788215Z 2025-09-07T08:06:59.5788321Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5788643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5788712Z layer_outputs = layer_module( 2025-09-07T08:06:59.5788925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5789000Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5789271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5789341Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5789604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5789670Z self_outputs = self.self( 2025-09-07T08:06:59.5789926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5790041Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5790364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5790513Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5790516Z 2025-09-07T08:06:59.5790615Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5790941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5791015Z layer_outputs = layer_module( 2025-09-07T08:06:59.5791222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5791304Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5791563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5791681Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5791966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5792050Z self_outputs = self.self( 2025-09-07T08:06:59.5792387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T08:06:59.5792578Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T08:06:59.5792617Z 2025-09-07T08:06:59.5792745Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5793258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5793343Z layer_outputs = layer_module( 2025-09-07T08:06:59.5816947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5817140Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5817490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5817583Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5817865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:06:59.5817980Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:59.5818267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:06:59.5818350Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:59.5818355Z 2025-09-07T08:06:59.5818448Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5818524Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5818597Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5818679Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5818787Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5819139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5819214Z layer_outputs = layer_module( 2025-09-07T08:06:59.5819439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5819522Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5819787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5819870Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5820132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5820213Z self_outputs = self.self( 2025-09-07T08:06:59.5820481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5820564Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5820568Z 2025-09-07T08:06:59.5820678Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5821018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5821100Z layer_outputs = layer_module( 2025-09-07T08:06:59.5821312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5821399Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5821668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5821741Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5822096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5822168Z self_outputs = self.self( 2025-09-07T08:06:59.5822443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5822601Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5822604Z 2025-09-07T08:06:59.5822707Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5823053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5823120Z layer_outputs = layer_module( 2025-09-07T08:06:59.5823339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5823415Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5823690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5823764Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5824027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5824105Z self_outputs = self.self( 2025-09-07T08:06:59.5824365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5824474Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5824795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5824974Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5824986Z 2025-09-07T08:06:59.5825087Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5825418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5825496Z layer_outputs = layer_module( 2025-09-07T08:06:59.5825710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5825791Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5826054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5826123Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5826393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5826458Z self_outputs = self.self( 2025-09-07T08:06:59.5826725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:06:59.5826799Z key_vectors = self.key(hidden_states) 2025-09-07T08:06:59.5826803Z 2025-09-07T08:06:59.5826906Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5827236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5827301Z layer_outputs = layer_module( 2025-09-07T08:06:59.5827516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5827588Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5827855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5827925Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5828524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5828599Z self_outputs = self.self( 2025-09-07T08:06:59.5828862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5828998Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5829319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5829477Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:06:59.5829741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:06:59.5829812Z hidden_states = hidden_states.view( 2025-09-07T08:06:59.5829818Z 2025-09-07T08:06:59.5829930Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5830255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5830336Z layer_outputs = layer_module( 2025-09-07T08:06:59.5830544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5830628Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5830890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5830961Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5831225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5831294Z self_outputs = self.self( 2025-09-07T08:06:59.5831563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5831660Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5831977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5832165Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5832168Z 2025-09-07T08:06:59.5832246Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5832331Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5832427Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5832764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5832834Z layer_outputs = layer_module( 2025-09-07T08:06:59.5833041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5833126Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5833389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5833467Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5833725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5833789Z self_outputs = self.self( 2025-09-07T08:06:59.5834058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:06:59.5834126Z attn_scores += diagonal_mask 2025-09-07T08:06:59.5834129Z 2025-09-07T08:06:59.5834261Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5834594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5834669Z layer_outputs = layer_module( 2025-09-07T08:06:59.5834906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5834981Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5835252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5835323Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5835589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5835655Z self_outputs = self.self( 2025-09-07T08:06:59.5835917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:06:59.5836002Z attn_probs = nn.functional.softmax( 2025-09-07T08:06:59.5836005Z 2025-09-07T08:06:59.5836103Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5836447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5836511Z layer_outputs = layer_module( 2025-09-07T08:06:59.5836724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5836796Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5837057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5837139Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5837400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5837472Z self_outputs = self.self( 2025-09-07T08:06:59.5837730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:06:59.5837813Z value_vectors = self.value(hidden_states) 2025-09-07T08:06:59.5837824Z 2025-09-07T08:06:59.5837919Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5838243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5838316Z layer_outputs = layer_module( 2025-09-07T08:06:59.5838524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5838608Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5838870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5838943Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5839220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5839285Z self_outputs = self.self( 2025-09-07T08:06:59.5839554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5839667Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5840001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5840201Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:06:59.5840388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5840489Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5840520Z 2025-09-07T08:06:59.5840618Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5840952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5841019Z layer_outputs = layer_module( 2025-09-07T08:06:59.5841224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5841306Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5841566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5841645Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5841906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5841980Z self_outputs = self.self( 2025-09-07T08:06:59.5842241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5842350Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5842684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5842815Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:06:59.5843117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:06:59.5843206Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:06:59.5843391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5843483Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5843489Z 2025-09-07T08:06:59.5843585Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5843917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5843987Z layer_outputs = layer_module( 2025-09-07T08:06:59.5844197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5844269Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5844531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5844611Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5844871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5844943Z self_outputs = self.self( 2025-09-07T08:06:59.5845205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5845322Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5845645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5845788Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5845791Z 2025-09-07T08:06:59.5845898Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5846254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5846333Z layer_outputs = layer_module( 2025-09-07T08:06:59.5846543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5846653Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5846915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5846985Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5847255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5847319Z self_outputs = self.self( 2025-09-07T08:06:59.5847590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5847699Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5848025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5848178Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5848181Z 2025-09-07T08:06:59.5848277Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5848612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5848679Z layer_outputs = layer_module( 2025-09-07T08:06:59.5848892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5848968Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5849227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5849307Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5849565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5849640Z self_outputs = self.self( 2025-09-07T08:06:59.5849895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T08:06:59.5850072Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T08:06:59.5850082Z 2025-09-07T08:06:59.5850178Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5850503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5850579Z layer_outputs = layer_module( 2025-09-07T08:06:59.5850781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5850868Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5851124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5851197Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5851463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:06:59.5851567Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:59.5851836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:06:59.5851944Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:59.5851947Z 2025-09-07T08:06:59.5852030Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5852102Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5852174Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5852288Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5852387Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5852728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5852796Z layer_outputs = layer_module( 2025-09-07T08:06:59.5853004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5853085Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5853352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5853431Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5853691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5853759Z self_outputs = self.self( 2025-09-07T08:06:59.5854032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5854111Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5854114Z 2025-09-07T08:06:59.5854219Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5854551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5854627Z layer_outputs = layer_module( 2025-09-07T08:06:59.5854840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5854914Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5855189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5855262Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5855529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5855596Z self_outputs = self.self( 2025-09-07T08:06:59.5855864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5855942Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5855946Z 2025-09-07T08:06:59.5856041Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5856373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5856438Z layer_outputs = layer_module( 2025-09-07T08:06:59.5856652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5856727Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5856986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5857064Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5857321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5857392Z self_outputs = self.self( 2025-09-07T08:06:59.5857683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5857787Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5858100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5858306Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5858309Z 2025-09-07T08:06:59.5858412Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5858735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5858805Z layer_outputs = layer_module( 2025-09-07T08:06:59.5859008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5859089Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5859352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5859421Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5859686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5859754Z self_outputs = self.self( 2025-09-07T08:06:59.5860022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:06:59.5860095Z key_vectors = self.key(hidden_states) 2025-09-07T08:06:59.5860098Z 2025-09-07T08:06:59.5860195Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5860527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5860593Z layer_outputs = layer_module( 2025-09-07T08:06:59.5860805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5860878Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5861146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5861217Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5861476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5861547Z self_outputs = self.self( 2025-09-07T08:06:59.5861807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5861910Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5862225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5862371Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:06:59.5862635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:06:59.5862707Z hidden_states = hidden_states.view( 2025-09-07T08:06:59.5862710Z 2025-09-07T08:06:59.5862812Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5863133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5863206Z layer_outputs = layer_module( 2025-09-07T08:06:59.5863411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5863510Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5863780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5863850Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5864144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5864211Z self_outputs = self.self( 2025-09-07T08:06:59.5864478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5864573Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5864885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5865067Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5865070Z 2025-09-07T08:06:59.5865144Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5865225Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5865322Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5865650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5865725Z layer_outputs = layer_module( 2025-09-07T08:06:59.5865934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5866013Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5866276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5866353Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5866617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5866682Z self_outputs = self.self( 2025-09-07T08:06:59.5866958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:06:59.5867028Z attn_scores += diagonal_mask 2025-09-07T08:06:59.5867031Z 2025-09-07T08:06:59.5867137Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5867471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5867535Z layer_outputs = layer_module( 2025-09-07T08:06:59.5867746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5867823Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5868094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5868164Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5868478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5868546Z self_outputs = self.self( 2025-09-07T08:06:59.5868804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:06:59.5868889Z attn_probs = nn.functional.softmax( 2025-09-07T08:06:59.5868892Z 2025-09-07T08:06:59.5868987Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5869323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5869459Z layer_outputs = layer_module( 2025-09-07T08:06:59.5869664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5869745Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5870056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5870134Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5870398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5870472Z self_outputs = self.self( 2025-09-07T08:06:59.5870737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:06:59.5870817Z value_vectors = self.value(hidden_states) 2025-09-07T08:06:59.5870820Z 2025-09-07T08:06:59.5870929Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5871256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5871331Z layer_outputs = layer_module( 2025-09-07T08:06:59.5871543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5871618Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5871890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5871960Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5872226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5872290Z self_outputs = self.self( 2025-09-07T08:06:59.5872557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5872667Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5872996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5873172Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:06:59.5873355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5873454Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5873458Z 2025-09-07T08:06:59.5873551Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5873885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5873954Z layer_outputs = layer_module( 2025-09-07T08:06:59.5874161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5874241Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5874504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5874583Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5874842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5874903Z self_outputs = self.self( 2025-09-07T08:06:59.5875168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5875304Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5875636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5875765Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:06:59.5876118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:06:59.5876203Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:06:59.5876384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5876484Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5876487Z 2025-09-07T08:06:59.5876583Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5876915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5876982Z layer_outputs = layer_module( 2025-09-07T08:06:59.5877196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5877272Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5877534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5877611Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5877871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5877942Z self_outputs = self.self( 2025-09-07T08:06:59.5878203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5878312Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5878644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5878788Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5878794Z 2025-09-07T08:06:59.5878900Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5879223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5879299Z layer_outputs = layer_module( 2025-09-07T08:06:59.5879508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5879582Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5879856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5879928Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5880195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5880262Z self_outputs = self.self( 2025-09-07T08:06:59.5880521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5880639Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5880963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5881111Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5881114Z 2025-09-07T08:06:59.5881243Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5881582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5881651Z layer_outputs = layer_module( 2025-09-07T08:06:59.5881889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5881971Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5882228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5882311Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5882570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5882642Z self_outputs = self.self( 2025-09-07T08:06:59.5882902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T08:06:59.5883076Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T08:06:59.5883079Z 2025-09-07T08:06:59.5883186Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5883514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5883588Z layer_outputs = layer_module( 2025-09-07T08:06:59.5883798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5883877Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5884137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5884208Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5884478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:06:59.5884582Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:59.5884853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:06:59.5884931Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:59.5884934Z 2025-09-07T08:06:59.5885007Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5885089Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5885162Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5885240Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5885338Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5885668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5885743Z layer_outputs = layer_module( 2025-09-07T08:06:59.5885948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5886031Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5886287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5886364Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5886620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5886682Z self_outputs = self.self( 2025-09-07T08:06:59.5886951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5887056Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5887059Z 2025-09-07T08:06:59.5887165Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5887494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5887601Z layer_outputs = layer_module( 2025-09-07T08:06:59.5887804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5887875Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5888143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5888212Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5888483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5888548Z self_outputs = self.self( 2025-09-07T08:06:59.5888807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5888891Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5888896Z 2025-09-07T08:06:59.5888992Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5889322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5889389Z layer_outputs = layer_module( 2025-09-07T08:06:59.5889602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5889675Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5889938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5890016Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5890274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5890347Z self_outputs = self.self( 2025-09-07T08:06:59.5890606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5890701Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5891025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5891197Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5891200Z 2025-09-07T08:06:59.5891305Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5891633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5891705Z layer_outputs = layer_module( 2025-09-07T08:06:59.5891919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5891999Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5892261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5892337Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5892600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5892662Z self_outputs = self.self( 2025-09-07T08:06:59.5892956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:06:59.5893031Z key_vectors = self.key(hidden_states) 2025-09-07T08:06:59.5893034Z 2025-09-07T08:06:59.5893139Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5893497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5893563Z layer_outputs = layer_module( 2025-09-07T08:06:59.5893777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5893849Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5894121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5894191Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5894463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5894529Z self_outputs = self.self( 2025-09-07T08:06:59.5894792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5894897Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5895218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5895372Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:06:59.5895639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:06:59.5895719Z hidden_states = hidden_states.view( 2025-09-07T08:06:59.5895721Z 2025-09-07T08:06:59.5895819Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5896149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5896227Z layer_outputs = layer_module( 2025-09-07T08:06:59.5896436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5896513Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5896778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5896847Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5897120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5897183Z self_outputs = self.self( 2025-09-07T08:06:59.5897460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5897556Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5897883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5898058Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5898061Z 2025-09-07T08:06:59.5898136Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5898219Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5898315Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5898656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5898747Z layer_outputs = layer_module( 2025-09-07T08:06:59.5898952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5899034Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5899291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5902725Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5902997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5903062Z self_outputs = self.self( 2025-09-07T08:06:59.5903335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:06:59.5903405Z attn_scores += diagonal_mask 2025-09-07T08:06:59.5903409Z 2025-09-07T08:06:59.5903519Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5903849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5903915Z layer_outputs = layer_module( 2025-09-07T08:06:59.5904129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5904236Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5904510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5904580Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5904841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5904917Z self_outputs = self.self( 2025-09-07T08:06:59.5905179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:06:59.5905263Z attn_probs = nn.functional.softmax( 2025-09-07T08:06:59.5905266Z 2025-09-07T08:06:59.5905363Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5905696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5905764Z layer_outputs = layer_module( 2025-09-07T08:06:59.5905973Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5906055Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5906315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5906391Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5906654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5906718Z self_outputs = self.self( 2025-09-07T08:06:59.5906985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:06:59.5907065Z value_vectors = self.value(hidden_states) 2025-09-07T08:06:59.5907069Z 2025-09-07T08:06:59.5907176Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5907507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5907583Z layer_outputs = layer_module( 2025-09-07T08:06:59.5907791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5907914Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5908187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5908259Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5908531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5908611Z self_outputs = self.self( 2025-09-07T08:06:59.5908913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5909030Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5909358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5909530Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:06:59.5909712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5909811Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5909815Z 2025-09-07T08:06:59.5909910Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5910238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5910316Z layer_outputs = layer_module( 2025-09-07T08:06:59.5910525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5910605Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5910866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5910945Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5911208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5911273Z self_outputs = self.self( 2025-09-07T08:06:59.5911539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5911650Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5911986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5912114Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:06:59.5912415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:06:59.5912510Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:06:59.5912691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5912790Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5912794Z 2025-09-07T08:06:59.5912889Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5913220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5913291Z layer_outputs = layer_module( 2025-09-07T08:06:59.5913496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5913573Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5913866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5913945Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5914214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5914277Z self_outputs = self.self( 2025-09-07T08:06:59.5914561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5914701Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5915030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5915170Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5915173Z 2025-09-07T08:06:59.5915275Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5915601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5915667Z layer_outputs = layer_module( 2025-09-07T08:06:59.5915879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5915954Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5916222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5916289Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5916556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5916619Z self_outputs = self.self( 2025-09-07T08:06:59.5916877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5916990Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5917309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5917456Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5917461Z 2025-09-07T08:06:59.5917555Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5917886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5917952Z layer_outputs = layer_module( 2025-09-07T08:06:59.5918158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5918237Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5918500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5918577Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5918838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5918905Z self_outputs = self.self( 2025-09-07T08:06:59.5919168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T08:06:59.5919338Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T08:06:59.5919342Z 2025-09-07T08:06:59.5919443Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5920111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5920191Z layer_outputs = layer_module( 2025-09-07T08:06:59.5920400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5920474Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5920760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5920850Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5921160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:06:59.5921266Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:59.5921535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:06:59.5921616Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:59.5921619Z 2025-09-07T08:06:59.5921696Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5921789Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5921861Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5921941Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5922039Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5922365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5922444Z layer_outputs = layer_module( 2025-09-07T08:06:59.5922651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5922728Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5922988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5923057Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5923325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5923390Z self_outputs = self.self( 2025-09-07T08:06:59.5923656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5923736Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5923740Z 2025-09-07T08:06:59.5923840Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5924165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5924233Z layer_outputs = layer_module( 2025-09-07T08:06:59.5924446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5924518Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5924788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5924859Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5925117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5925190Z self_outputs = self.self( 2025-09-07T08:06:59.5925453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5925537Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5925540Z 2025-09-07T08:06:59.5925634Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5925991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5926058Z layer_outputs = layer_module( 2025-09-07T08:06:59.5926261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5926356Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5926616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5926717Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5926977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5927040Z self_outputs = self.self( 2025-09-07T08:06:59.5927307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5927405Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5927728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5927903Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5927906Z 2025-09-07T08:06:59.5928010Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5928335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5928399Z layer_outputs = layer_module( 2025-09-07T08:06:59.5928610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5928682Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5928949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5929019Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5929282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5929348Z self_outputs = self.self( 2025-09-07T08:06:59.5929604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:06:59.5929685Z key_vectors = self.key(hidden_states) 2025-09-07T08:06:59.5929688Z 2025-09-07T08:06:59.5929783Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5930111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5930179Z layer_outputs = layer_module( 2025-09-07T08:06:59.5930383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5930461Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5930717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5930793Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5931050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5931120Z self_outputs = self.self( 2025-09-07T08:06:59.5931374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5931469Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5931816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5931965Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:06:59.5932228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:06:59.5932311Z hidden_states = hidden_states.view( 2025-09-07T08:06:59.5932328Z 2025-09-07T08:06:59.5932429Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5932759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5932823Z layer_outputs = layer_module( 2025-09-07T08:06:59.5933038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5933111Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5933381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5933450Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5933709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5933783Z self_outputs = self.self( 2025-09-07T08:06:59.5934042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5934143Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5934454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5934631Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5934634Z 2025-09-07T08:06:59.5934708Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5934782Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5934886Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5935216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5935291Z layer_outputs = layer_module( 2025-09-07T08:06:59.5935498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5935570Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5935838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5935908Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5936177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5936241Z self_outputs = self.self( 2025-09-07T08:06:59.5936507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:06:59.5936578Z attn_scores += diagonal_mask 2025-09-07T08:06:59.5936581Z 2025-09-07T08:06:59.5936680Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5937018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5937084Z layer_outputs = layer_module( 2025-09-07T08:06:59.5937299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5937371Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5937657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5937735Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5937992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5938080Z self_outputs = self.self( 2025-09-07T08:06:59.5938352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:06:59.5938433Z attn_probs = nn.functional.softmax( 2025-09-07T08:06:59.5938436Z 2025-09-07T08:06:59.5938530Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5938854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5938932Z layer_outputs = layer_module( 2025-09-07T08:06:59.5939136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5939220Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5939482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5939555Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5939826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5939893Z self_outputs = self.self( 2025-09-07T08:06:59.5940161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:06:59.5940242Z value_vectors = self.value(hidden_states) 2025-09-07T08:06:59.5940245Z 2025-09-07T08:06:59.5940351Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5940680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5940750Z layer_outputs = layer_module( 2025-09-07T08:06:59.5940966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5941044Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5941313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5941385Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5941651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5941720Z self_outputs = self.self( 2025-09-07T08:06:59.5941983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5942102Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5942429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5942600Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:06:59.5942783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5942876Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5942885Z 2025-09-07T08:06:59.5942982Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5943340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5943414Z layer_outputs = layer_module( 2025-09-07T08:06:59.5943618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5943698Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5943973Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5944061Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5944330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5944395Z self_outputs = self.self( 2025-09-07T08:06:59.5944658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5944765Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5945088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5945221Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:06:59.5945517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:06:59.5945611Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:06:59.5945789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5945885Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5945888Z 2025-09-07T08:06:59.5945983Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5946308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5946381Z layer_outputs = layer_module( 2025-09-07T08:06:59.5946586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5946664Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5946929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5947006Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5947268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5947330Z self_outputs = self.self( 2025-09-07T08:06:59.5947599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5947705Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5948039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5948180Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5948185Z 2025-09-07T08:06:59.5948290Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5948617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5948684Z layer_outputs = layer_module( 2025-09-07T08:06:59.5948898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5948970Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5949269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5949341Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5949598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5949695Z self_outputs = self.self( 2025-09-07T08:06:59.5949953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5950080Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5950409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5950553Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5950556Z 2025-09-07T08:06:59.5950655Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5950981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5951056Z layer_outputs = layer_module( 2025-09-07T08:06:59.5951263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5951345Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5951609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5951679Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5951946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5952011Z self_outputs = self.self( 2025-09-07T08:06:59.5952281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T08:06:59.5952458Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T08:06:59.5952461Z 2025-09-07T08:06:59.5952564Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5952890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5952958Z layer_outputs = layer_module( 2025-09-07T08:06:59.5953169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5953242Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5953506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5953577Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5953844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:06:59.5953947Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:59.5954207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:06:59.5954294Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:59.5954297Z 2025-09-07T08:06:59.5954372Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5954452Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5954525Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5954594Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5954699Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5955052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5955128Z layer_outputs = layer_module( 2025-09-07T08:06:59.5955332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5955420Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5955685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5955770Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5956041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5956102Z self_outputs = self.self( 2025-09-07T08:06:59.5956370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5956451Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5956454Z 2025-09-07T08:06:59.5956550Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5956885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5956953Z layer_outputs = layer_module( 2025-09-07T08:06:59.5957164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5957236Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5957502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5957572Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5957836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5957907Z self_outputs = self.self( 2025-09-07T08:06:59.5958167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5958250Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5958254Z 2025-09-07T08:06:59.5958347Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5958672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5958743Z layer_outputs = layer_module( 2025-09-07T08:06:59.5958950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5959030Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5959297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5959375Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5959639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5959704Z self_outputs = self.self( 2025-09-07T08:06:59.5959971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5960066Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5960389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5960562Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5960566Z 2025-09-07T08:06:59.5960659Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5961028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5961097Z layer_outputs = layer_module( 2025-09-07T08:06:59.5961310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5961396Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5961680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5961748Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5962007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5962077Z self_outputs = self.self( 2025-09-07T08:06:59.5962337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:06:59.5962417Z key_vectors = self.key(hidden_states) 2025-09-07T08:06:59.5962420Z 2025-09-07T08:06:59.5962514Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5962847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5962917Z layer_outputs = layer_module( 2025-09-07T08:06:59.5963122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5963203Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5963462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5963537Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5963799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5963863Z self_outputs = self.self( 2025-09-07T08:06:59.5964130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5964227Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5964546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5964694Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:06:59.5964958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:06:59.5965029Z hidden_states = hidden_states.view( 2025-09-07T08:06:59.5965031Z 2025-09-07T08:06:59.5965131Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5965464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5965532Z layer_outputs = layer_module( 2025-09-07T08:06:59.5965745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5965821Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5966084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5966161Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5966419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5966491Z self_outputs = self.self( 2025-09-07T08:06:59.5966781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5966884Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5967196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5967384Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5967402Z 2025-09-07T08:06:59.5967486Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5967561Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5967662Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5967988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5968061Z layer_outputs = layer_module( 2025-09-07T08:06:59.5968313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5968386Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5968656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5968727Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5968998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5969062Z self_outputs = self.self( 2025-09-07T08:06:59.5969323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:06:59.5969402Z attn_scores += diagonal_mask 2025-09-07T08:06:59.5969404Z 2025-09-07T08:06:59.5969503Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5969839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5969905Z layer_outputs = layer_module( 2025-09-07T08:06:59.5970120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5970193Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5970453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5970529Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5970791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5970862Z self_outputs = self.self( 2025-09-07T08:06:59.5971122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:06:59.5971196Z attn_probs = nn.functional.softmax( 2025-09-07T08:06:59.5971199Z 2025-09-07T08:06:59.5971302Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5971631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5971704Z layer_outputs = layer_module( 2025-09-07T08:06:59.5971909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5971986Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5972247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5972318Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5972647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5972711Z self_outputs = self.self( 2025-09-07T08:06:59.5972978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:06:59.5973081Z value_vectors = self.value(hidden_states) 2025-09-07T08:06:59.5973108Z 2025-09-07T08:06:59.5973204Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5973534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5973600Z layer_outputs = layer_module( 2025-09-07T08:06:59.5973808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5973879Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5974148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5974217Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5974476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5974552Z self_outputs = self.self( 2025-09-07T08:06:59.5974814Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5974931Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5975258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5975429Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:06:59.5975611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5975704Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5975707Z 2025-09-07T08:06:59.5975810Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5976136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5976213Z layer_outputs = layer_module( 2025-09-07T08:06:59.5976415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5976488Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5976753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5976825Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5977091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5977153Z self_outputs = self.self( 2025-09-07T08:06:59.5977418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5977526Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5977852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5977985Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:06:59.5978276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:06:59.5978399Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:06:59.5978580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.5978674Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.5978685Z 2025-09-07T08:06:59.5978795Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5979120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5979211Z layer_outputs = layer_module( 2025-09-07T08:06:59.5979416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5979496Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5979757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5979830Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5980096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5980159Z self_outputs = self.self( 2025-09-07T08:06:59.5980430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5980538Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5980875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5981016Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5981019Z 2025-09-07T08:06:59.5981115Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5981453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5981523Z layer_outputs = layer_module( 2025-09-07T08:06:59.5981736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5981809Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5982074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5982151Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5982413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5982487Z self_outputs = self.self( 2025-09-07T08:06:59.5982748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.5982869Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.5983196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.5983338Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.5983341Z 2025-09-07T08:06:59.5983446Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5983771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5983844Z layer_outputs = layer_module( 2025-09-07T08:06:59.5984050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5984127Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5984416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5984488Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5984757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5984838Z self_outputs = self.self( 2025-09-07T08:06:59.5985120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T08:06:59.5985291Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T08:06:59.5985295Z 2025-09-07T08:06:59.5985397Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5985724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5985792Z layer_outputs = layer_module( 2025-09-07T08:06:59.5986004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5986075Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5986345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5986416Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5986672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:06:59.5986781Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:59.5987041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:06:59.5987131Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:59.5987134Z 2025-09-07T08:06:59.5987210Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5987291Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5987362Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5987430Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.5987533Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5987857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5987933Z layer_outputs = layer_module( 2025-09-07T08:06:59.5988138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5988210Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5988476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5988545Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5988806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5988870Z self_outputs = self.self( 2025-09-07T08:06:59.5989128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5989213Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5989216Z 2025-09-07T08:06:59.5989310Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5989642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5989708Z layer_outputs = layer_module( 2025-09-07T08:06:59.5989946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5990016Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5990272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5990366Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5990622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5990710Z self_outputs = self.self( 2025-09-07T08:06:59.5990966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:06:59.5991041Z query_vectors = self.query(hidden_states) 2025-09-07T08:06:59.5991052Z 2025-09-07T08:06:59.5991146Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5991469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5991542Z layer_outputs = layer_module( 2025-09-07T08:06:59.5991750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5991832Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5992090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5992160Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5992423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5992486Z self_outputs = self.self( 2025-09-07T08:06:59.5992748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5992843Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5993157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5993332Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.5993337Z 2025-09-07T08:06:59.5993431Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5993764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5993832Z layer_outputs = layer_module( 2025-09-07T08:06:59.5994046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5994119Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5994379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5994455Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5994713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5994787Z self_outputs = self.self( 2025-09-07T08:06:59.5995043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:06:59.5995125Z key_vectors = self.key(hidden_states) 2025-09-07T08:06:59.5995128Z 2025-09-07T08:06:59.5995223Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5995545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5995618Z layer_outputs = layer_module( 2025-09-07T08:06:59.5995863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5995942Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5996202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5996283Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5996564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5996627Z self_outputs = self.self( 2025-09-07T08:06:59.5996891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5996983Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5997303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.5997447Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:06:59.5997706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:06:59.5997784Z hidden_states = hidden_states.view( 2025-09-07T08:06:59.5997789Z 2025-09-07T08:06:59.5997884Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.5998211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.5998277Z layer_outputs = layer_module( 2025-09-07T08:06:59.5998485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.5998557Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.5998818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.5998893Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.5999150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.5999221Z self_outputs = self.self( 2025-09-07T08:06:59.5999484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:06:59.5999575Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:06:59.5999894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:06:59.6000066Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:06:59.6000069Z 2025-09-07T08:06:59.6000148Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.6000221Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.6000324Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.6000649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.6000719Z layer_outputs = layer_module( 2025-09-07T08:06:59.6000931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.6001002Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.6001268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.6001336Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.6001622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.6001693Z self_outputs = self.self( 2025-09-07T08:06:59.6001952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:06:59.6002042Z attn_scores += diagonal_mask 2025-09-07T08:06:59.6002045Z 2025-09-07T08:06:59.6002167Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.6002502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.6002567Z layer_outputs = layer_module( 2025-09-07T08:06:59.6002775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.6002855Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.6003116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.6003193Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.6003456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.6003520Z self_outputs = self.self( 2025-09-07T08:06:59.6003789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:06:59.6003861Z attn_probs = nn.functional.softmax( 2025-09-07T08:06:59.6003864Z 2025-09-07T08:06:59.6003967Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.6004291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.6004368Z layer_outputs = layer_module( 2025-09-07T08:06:59.6004567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.6004638Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.6004904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.6004973Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.6005239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.6005302Z self_outputs = self.self( 2025-09-07T08:06:59.6005559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:06:59.6005647Z value_vectors = self.value(hidden_states) 2025-09-07T08:06:59.6005650Z 2025-09-07T08:06:59.6005745Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.6006079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.6006146Z layer_outputs = layer_module( 2025-09-07T08:06:59.6006358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.6006427Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.6006689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.6006764Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.6007022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.6007095Z self_outputs = self.self( 2025-09-07T08:06:59.6007383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.6007498Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.6007822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.6008000Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:06:59.6008223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.6008315Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.6008319Z 2025-09-07T08:06:59.6008418Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.6008747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.6008820Z layer_outputs = layer_module( 2025-09-07T08:06:59.6009025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.6009095Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.6009363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.6009433Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.6009700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.6009763Z self_outputs = self.self( 2025-09-07T08:06:59.6010020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.6010134Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.6010458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.6010586Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:06:59.6010880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:06:59.6010974Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:06:59.6011150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:06:59.6011240Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:06:59.6011243Z 2025-09-07T08:06:59.6011345Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.6011674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.6011748Z layer_outputs = layer_module( 2025-09-07T08:06:59.6011953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.6012025Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.6012294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.6012365Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.6012635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.6012698Z self_outputs = self.self( 2025-09-07T08:06:59.6012965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.6013099Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.6013427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.6013573Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.6013593Z 2025-09-07T08:06:59.6013690Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.6014042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.6014107Z layer_outputs = layer_module( 2025-09-07T08:06:59.6014318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.6014389Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.6014649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.6014726Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.6014984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.6015055Z self_outputs = self.self( 2025-09-07T08:06:59.6015309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:06:59.6015413Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:06:59.6015741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:06:59.6015877Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:06:59.6015881Z 2025-09-07T08:06:59.6015984Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.6016310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.6016384Z layer_outputs = layer_module( 2025-09-07T08:06:59.6016587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.6016659Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.6016925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.6016994Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.6017257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:06:59.6017319Z self_outputs = self.self( 2025-09-07T08:06:59.6017586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T08:06:59.6017756Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T08:06:59.6017759Z 2025-09-07T08:06:59.6017855Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:59.6018186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:06:59.6018253Z layer_outputs = layer_module( 2025-09-07T08:06:59.6018465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:59.6018536Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:59.6018791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:06:59.6018895Z self_attn_outputs = self.attention( 2025-09-07T08:06:59.6019154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:06:59.6019265Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:59.6019541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:06:59.6019638Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:59.6019641Z 2025-09-07T08:06:59.6019714Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.6019785Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.6019863Z cudagraph partition due to non gpu ops 2025-09-07T08:06:59.6019932Z cudagraph partition due to non gpu ops 2025-09-07T08:07:20.3067830Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:20.3068550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1716, in torch_dynamo_resume_in_forward_at_1703 2025-09-07T08:07:20.3069058Z prediction_scores = self.lm_head(sequence_output) 2025-09-07T08:07:20.3069467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1333, in forward 2025-09-07T08:07:20.3069874Z x = self.dense(features) 2025-09-07T08:07:20.3070000Z 2025-09-07T08:07:20.3070084Z cudagraph partition due to non gpu ops 2025-09-07T08:07:20.3070297Z cudagraph partition due to non gpu ops 2025-09-07T08:07:20.3070493Z cudagraph partition due to non gpu ops 2025-09-07T08:07:20.3070681Z cudagraph partition due to non gpu ops 2025-09-07T08:07:23.2819363Z pass 2025-09-07T08:07:23.2819784Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:07:25.7030854Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:07:25.7032265Z import pynvml # type: ignore[import] 2025-09-07T08:07:27.4245181Z 2025-09-07T08:07:28.4995694Z loading model: 0it [00:00, ?it/s]Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T08:07:28.4998187Z WARNING:transformers.models.reformer.modeling_reformer:Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T08:07:28.8994296Z 2025-09-07T08:07:28.8994856Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:07:28.8995341Z cpu eval hf_Reformer 2025-09-07T08:07:29.2451000Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:07:30.0054969Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:07:30.7654369Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:07:39.5106795Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:39.5107381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2095, in forward 2025-09-07T08:07:39.5107861Z least_common_mult_chunk_length = _get_least_common_mult_chunk_len(self.config) 2025-09-07T08:07:39.5108398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 194, in _get_least_common_mult_chunk_len 2025-09-07T08:07:39.5109233Z return np.lcm(config.lsh_attn_chunk_length, config.local_attn_chunk_length) 2025-09-07T08:07:39.5109469Z 2025-09-07T08:07:39.5109576Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:39.5110005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2099, in forward 2025-09-07T08:07:39.5110497Z input_shape[-1] % least_common_mult_chunk_length != 0 2025-09-07T08:07:39.5110651Z 2025-09-07T08:07:39.8803049Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T08:07:39.8804028Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T08:07:39.8804527Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] or: 2025-09-07T08:07:39.8805009Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T08:07:39.8805592Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] to include these operations in the captured graph. 2025-09-07T08:07:39.8806066Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T08:07:39.8806526Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] Graph break: from user code at: 2025-09-07T08:07:39.8807349Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2132, in torch_dynamo_resume_in_forward_at_2104 2025-09-07T08:07:39.8808194Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] embedding_output = self.embeddings( 2025-09-07T08:07:39.8808931Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 369, in forward 2025-09-07T08:07:39.8809700Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] position_embeddings = self.position_embeddings(position_ids) 2025-09-07T08:07:39.8810476Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 293, in forward 2025-09-07T08:07:39.8811210Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] max_position_id = position_ids.max().item() 2025-09-07T08:07:39.8811674Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T08:07:39.8812044Z W0907 08:07:39.879483 45937 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T08:07:40.0194048Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:40.0194545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 351, in forward 2025-09-07T08:07:40.0194947Z position_ids = torch.arange( 2025-09-07T08:07:40.0195082Z 2025-09-07T08:07:40.0195195Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:40.0195617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 357, in forward 2025-09-07T08:07:40.0196026Z inputs_embeds = self.word_embeddings(input_ids) 2025-09-07T08:07:40.0196182Z 2025-09-07T08:07:40.9062480Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:40.9063013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 293, in forward 2025-09-07T08:07:40.9063434Z max_position_id = position_ids.max().item() 2025-09-07T08:07:40.9063576Z 2025-09-07T08:07:42.2582118Z cudagraph partition due to non gpu ops 2025-09-07T08:07:42.2582442Z cudagraph partition due to non gpu ops 2025-09-07T08:07:42.2582678Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:42.2583208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 297, in torch_dynamo_resume_in_forward_at_293 2025-09-07T08:07:42.2583747Z position_encodings = torch.cat( 2025-09-07T08:07:42.2583883Z 2025-09-07T08:07:42.2584046Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:42.2584528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 303, in torch_dynamo_resume_in_forward_at_293 2025-09-07T08:07:42.2584983Z position_encodings = torch.cat( 2025-09-07T08:07:42.2585110Z 2025-09-07T08:07:43.0932137Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:43.0932751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 370, in torch_dynamo_resume_in_forward_at_369 2025-09-07T08:07:43.0933250Z embeddings = embeddings + position_embeddings 2025-09-07T08:07:43.0933409Z 2025-09-07T08:07:47.3317347Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3318002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3318687Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3319151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3319630Z layer_outputs = layer( 2025-09-07T08:07:47.3320074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3320562Z attn_outputs = self.attention( 2025-09-07T08:07:47.3321031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3321486Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3321930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3322516Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3323113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 749, in _hash_vectors 2025-09-07T08:07:47.3323709Z random_rotations = torch.randn(rotations_shape, device=vectors.device, dtype=vectors.dtype) 2025-09-07T08:07:47.3324013Z 2025-09-07T08:07:47.3324147Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3324769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3325401Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3325893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3326334Z layer_outputs = layer( 2025-09-07T08:07:47.3326799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3327262Z attn_outputs = self.attention( 2025-09-07T08:07:47.3327744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3328161Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3328575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:07:47.3329484Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:07:47.3330127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:07:47.3330697Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T08:07:47.3330926Z 2025-09-07T08:07:47.3331045Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3331590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3332096Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3332535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3332920Z layer_outputs = layer( 2025-09-07T08:07:47.3333297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3333727Z attn_outputs = self.attention( 2025-09-07T08:07:47.3334127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3334561Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3334969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:07:47.3335476Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:07:47.3336100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:07:47.3336628Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:07:47.3337104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 180, in _stable_argsort 2025-09-07T08:07:47.3337607Z scale_offset = torch.arange(vector.shape[dim], device=vector.device).view(1, 1, -1) 2025-09-07T08:07:47.3337849Z 2025-09-07T08:07:47.3337952Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3338474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3338998Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3339438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3339869Z layer_outputs = layer( 2025-09-07T08:07:47.3340277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3340724Z attn_outputs = self.attention( 2025-09-07T08:07:47.3341142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3341586Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3342017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3342500Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3342958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 749, in _hash_vectors 2025-09-07T08:07:47.3343447Z random_rotations = torch.randn(rotations_shape, device=vectors.device, dtype=vectors.dtype) 2025-09-07T08:07:47.3343681Z 2025-09-07T08:07:47.3343780Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3344282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3344742Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3345126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3345517Z layer_outputs = layer( 2025-09-07T08:07:47.3345894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3346270Z attn_outputs = self.attention( 2025-09-07T08:07:47.3346640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3347024Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3347418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:07:47.3347896Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:07:47.3348452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:07:47.3348954Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T08:07:47.3349119Z 2025-09-07T08:07:47.3349217Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3349680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3350130Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3350518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3350887Z layer_outputs = layer( 2025-09-07T08:07:47.3351233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3351606Z attn_outputs = self.attention( 2025-09-07T08:07:47.3351978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3352369Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3352758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3353189Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3353643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 749, in _hash_vectors 2025-09-07T08:07:47.3354141Z random_rotations = torch.randn(rotations_shape, device=vectors.device, dtype=vectors.dtype) 2025-09-07T08:07:47.3354366Z 2025-09-07T08:07:47.3354475Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3354941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3355393Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3355784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3356157Z layer_outputs = layer( 2025-09-07T08:07:47.3356513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3356890Z attn_outputs = self.attention( 2025-09-07T08:07:47.3357282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3357682Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3358076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:07:47.3358555Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:07:47.3359130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:07:47.3359642Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T08:07:47.3359815Z 2025-09-07T08:07:47.3359893Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3360103Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3360302Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3360492Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3360695Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3360888Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3361111Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3361582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3362039Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3362438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3362811Z layer_outputs = layer( 2025-09-07T08:07:47.3363170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3363538Z attn_outputs = self.attention( 2025-09-07T08:07:47.3363916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3364307Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3364699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:07:47.3365180Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3365665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3366129Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3366334Z 2025-09-07T08:07:47.3366411Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3366615Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3366842Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3367305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3367765Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3368161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3368628Z layer_outputs = layer( 2025-09-07T08:07:47.3368995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3369376Z attn_outputs = self.attention( 2025-09-07T08:07:47.3369749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3370140Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3370614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:07:47.3371093Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3371576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:07:47.3371999Z return torch.cat(slices, dim=3) 2025-09-07T08:07:47.3372131Z 2025-09-07T08:07:47.3372238Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3372738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3373197Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3373581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3373959Z layer_outputs = layer( 2025-09-07T08:07:47.3374312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3374685Z attn_outputs = self.attention( 2025-09-07T08:07:47.3375048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3375440Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3375834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:07:47.3376283Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:07:47.3376472Z 2025-09-07T08:07:47.3376577Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3377031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3377487Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3377877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3378244Z layer_outputs = layer( 2025-09-07T08:07:47.3378600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3378968Z attn_outputs = self.attention( 2025-09-07T08:07:47.3379343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3379731Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3380120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:07:47.3380562Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:07:47.3380750Z 2025-09-07T08:07:47.3380826Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3381026Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3381223Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3381416Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3381627Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3382099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3382556Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3382946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3383315Z layer_outputs = layer( 2025-09-07T08:07:47.3383662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3384067Z attn_outputs = self.attention( 2025-09-07T08:07:47.3384436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3384827Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3385238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T08:07:47.3385727Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3386223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3386680Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3386868Z 2025-09-07T08:07:47.3386952Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3387152Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3387369Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3387832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3388287Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3388682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3389048Z layer_outputs = layer( 2025-09-07T08:07:47.3389403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3389779Z attn_outputs = self.attention( 2025-09-07T08:07:47.3390147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3390537Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3390923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:07:47.3391333Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:07:47.3391500Z 2025-09-07T08:07:47.3391599Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3392062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3392517Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3392897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3393266Z layer_outputs = layer( 2025-09-07T08:07:47.3393626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3394000Z attn_outputs = self.attention( 2025-09-07T08:07:47.3394368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3394748Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3395139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:07:47.3395553Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:07:47.3395708Z 2025-09-07T08:07:47.3395815Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3396276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3396720Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3398636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3399031Z layer_outputs = layer( 2025-09-07T08:07:47.3399396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3399801Z attn_outputs = self.attention( 2025-09-07T08:07:47.3400164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3400610Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3401005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T08:07:47.3401505Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:07:47.3402047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:07:47.3402529Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:07:47.3402719Z 2025-09-07T08:07:47.3402796Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3403006Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3403208Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3403399Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3403596Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3403793Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3403987Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3404171Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3404394Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3404863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3405324Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3405723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3406089Z layer_outputs = layer( 2025-09-07T08:07:47.3406454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3406837Z attn_outputs = self.attention( 2025-09-07T08:07:47.3407212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3407609Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3407994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3408437Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3408895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 749, in _hash_vectors 2025-09-07T08:07:47.3409405Z random_rotations = torch.randn(rotations_shape, device=vectors.device, dtype=vectors.dtype) 2025-09-07T08:07:47.3409636Z 2025-09-07T08:07:47.3409742Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3410204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3410673Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3411065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3411440Z layer_outputs = layer( 2025-09-07T08:07:47.3411859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3412225Z attn_outputs = self.attention( 2025-09-07T08:07:47.3412590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3413002Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3413390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3413839Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3414293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T08:07:47.3414753Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T08:07:47.3414955Z 2025-09-07T08:07:47.3415054Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3415533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3415991Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3416377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3416744Z layer_outputs = layer( 2025-09-07T08:07:47.3417100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3417476Z attn_outputs = self.attention( 2025-09-07T08:07:47.3417840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3418233Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3418628Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3419065Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3419512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T08:07:47.3419957Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T08:07:47.3420146Z 2025-09-07T08:07:47.3420242Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3420704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3421164Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3421551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3421913Z layer_outputs = layer( 2025-09-07T08:07:47.3422263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3422638Z attn_outputs = self.attention( 2025-09-07T08:07:47.3423007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3423397Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3423778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3424213Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3424660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T08:07:47.3425112Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T08:07:47.3425257Z 2025-09-07T08:07:47.3425365Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3425828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3426300Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3426691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3427083Z layer_outputs = layer( 2025-09-07T08:07:47.3427442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3427811Z attn_outputs = self.attention( 2025-09-07T08:07:47.3428189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3428585Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3428976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:07:47.3429463Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:07:47.3430030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:07:47.3430522Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:07:47.3430949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T08:07:47.3431422Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T08:07:47.3431619Z 2025-09-07T08:07:47.3431705Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3431934Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3432403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3432867Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3433262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3433638Z layer_outputs = layer( 2025-09-07T08:07:47.3433995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3434377Z attn_outputs = self.attention( 2025-09-07T08:07:47.3434752Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3435146Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3435534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:07:47.3436017Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:07:47.3436583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:07:47.3437073Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:07:47.3437497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T08:07:47.3437900Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T08:07:47.3438044Z 2025-09-07T08:07:47.3438127Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3438335Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3438591Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3439058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3439511Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3439918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3440306Z layer_outputs = layer( 2025-09-07T08:07:47.3440662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3441033Z attn_outputs = self.attention( 2025-09-07T08:07:47.3441395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3441789Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3442180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3442589Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3442980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:07:47.3443480Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3443992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3444451Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3444643Z 2025-09-07T08:07:47.3444724Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3444929Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3445145Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3445608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3446061Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3446453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3446819Z layer_outputs = layer( 2025-09-07T08:07:47.3447177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3447551Z attn_outputs = self.attention( 2025-09-07T08:07:47.3447920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3448308Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3448690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3449091Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3449487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:07:47.3449981Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3450493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:07:47.3450877Z return torch.cat(slices, dim=3) 2025-09-07T08:07:47.3451005Z 2025-09-07T08:07:47.3451079Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3451277Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3451470Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3451709Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3452174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3452629Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3453040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3453426Z layer_outputs = layer( 2025-09-07T08:07:47.3453775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3454146Z attn_outputs = self.attention( 2025-09-07T08:07:47.3454510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3454900Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3455290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3455685Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3456082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:07:47.3456553Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3457037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3457488Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3457678Z 2025-09-07T08:07:47.3457753Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3457976Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3458441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3458898Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3459288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3459653Z layer_outputs = layer( 2025-09-07T08:07:47.3460011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3460388Z attn_outputs = self.attention( 2025-09-07T08:07:47.3460760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3461142Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3461533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3461931Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3462326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:07:47.3462797Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3463277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:07:47.3463673Z return torch.cat(slices, dim=3) 2025-09-07T08:07:47.3463798Z 2025-09-07T08:07:47.3463894Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3464356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3464836Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3465221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3465590Z layer_outputs = layer( 2025-09-07T08:07:47.3465943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3466342Z attn_outputs = self.attention( 2025-09-07T08:07:47.3466772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3467157Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3467549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T08:07:47.3468046Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T08:07:47.3468647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T08:07:47.3469084Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T08:07:47.3469223Z 2025-09-07T08:07:47.3469323Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3469802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3470263Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3470657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3471031Z layer_outputs = layer( 2025-09-07T08:07:47.3471379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3471760Z attn_outputs = self.attention( 2025-09-07T08:07:47.3472134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3472526Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3472916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3473310Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3473711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T08:07:47.3474162Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:07:47.3474352Z 2025-09-07T08:07:47.3474462Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3474929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3475378Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3475774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3476147Z layer_outputs = layer( 2025-09-07T08:07:47.3476504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3476886Z attn_outputs = self.attention( 2025-09-07T08:07:47.3477248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3477642Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3478030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3478491Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3478883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T08:07:47.3479369Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3479891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3480383Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3480574Z 2025-09-07T08:07:47.3480663Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3480895Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3481370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3481847Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3482246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3482623Z layer_outputs = layer( 2025-09-07T08:07:47.3482979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3483367Z attn_outputs = self.attention( 2025-09-07T08:07:47.3483745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3484144Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3484540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3484940Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3485348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:07:47.3485764Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:07:47.3485923Z 2025-09-07T08:07:47.3486030Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3486501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3486958Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3487354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3487727Z layer_outputs = layer( 2025-09-07T08:07:47.3488085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3488470Z attn_outputs = self.attention( 2025-09-07T08:07:47.3488837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3489234Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3489624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3490027Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3490423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:07:47.3490838Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:07:47.3491000Z 2025-09-07T08:07:47.3491080Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3491311Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3491813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3492267Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3492661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3493053Z layer_outputs = layer( 2025-09-07T08:07:47.3493415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3493805Z attn_outputs = self.attention( 2025-09-07T08:07:47.3494166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3494556Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3494948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T08:07:47.3495447Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:07:47.3495986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:07:47.3496463Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:07:47.3496656Z 2025-09-07T08:07:47.3496733Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3496937Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3497133Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3497329Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3497519Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3497716Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3497906Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3498093Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3498278Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3498473Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3498667Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3498855Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3499068Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3499537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3499997Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3500391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3500755Z layer_outputs = layer( 2025-09-07T08:07:47.3501112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3501492Z attn_outputs = self.attention( 2025-09-07T08:07:47.3501866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3502258Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3502646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:07:47.3503121Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3503609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3504070Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3504259Z 2025-09-07T08:07:47.3504340Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3504530Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3504782Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3505250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3505721Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3506112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3506495Z layer_outputs = layer( 2025-09-07T08:07:47.3506854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3507234Z attn_outputs = self.attention( 2025-09-07T08:07:47.3507605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3508001Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3508387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:07:47.3508862Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3509356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:07:47.3509751Z return torch.cat(slices, dim=3) 2025-09-07T08:07:47.3509872Z 2025-09-07T08:07:47.3509976Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3510433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3510891Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3511282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3511655Z layer_outputs = layer( 2025-09-07T08:07:47.3512009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3512393Z attn_outputs = self.attention( 2025-09-07T08:07:47.3512767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3513161Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3513554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:07:47.3513993Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:07:47.3514192Z 2025-09-07T08:07:47.3514289Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3514763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3515217Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3515603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3515969Z layer_outputs = layer( 2025-09-07T08:07:47.3516330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3516707Z attn_outputs = self.attention( 2025-09-07T08:07:47.3517076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3517468Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3517879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:07:47.3518328Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:07:47.3518525Z 2025-09-07T08:07:47.3518599Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3518799Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3519016Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3519214Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3519455Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3519935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3520394Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3520781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3521154Z layer_outputs = layer( 2025-09-07T08:07:47.3521514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3521895Z attn_outputs = self.attention( 2025-09-07T08:07:47.3522265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3522653Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3523047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T08:07:47.3523530Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3524030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3524492Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3524685Z 2025-09-07T08:07:47.3524761Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3525085Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3525310Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3525779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3526240Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3526626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3527000Z layer_outputs = layer( 2025-09-07T08:07:47.3527356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3527737Z attn_outputs = self.attention( 2025-09-07T08:07:47.3528108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3528505Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3528896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:07:47.3529317Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:07:47.3529474Z 2025-09-07T08:07:47.3529584Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3530044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3530500Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3530890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3531291Z layer_outputs = layer( 2025-09-07T08:07:47.3531651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3532024Z attn_outputs = self.attention( 2025-09-07T08:07:47.3532400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3532806Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3533215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:07:47.3533631Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:07:47.3533789Z 2025-09-07T08:07:47.3533885Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3534357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3534810Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3535195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3535567Z layer_outputs = layer( 2025-09-07T08:07:47.3535916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3536298Z attn_outputs = self.attention( 2025-09-07T08:07:47.3536671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3537063Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3537446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T08:07:47.3537949Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:07:47.3538491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:07:47.3538967Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:07:47.3539151Z 2025-09-07T08:07:47.3539234Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3539435Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3539624Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3539819Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3540014Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3540207Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3540390Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3540580Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3540800Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3541270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3541721Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3542113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3542484Z layer_outputs = layer( 2025-09-07T08:07:47.3542846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3543221Z attn_outputs = self.attention( 2025-09-07T08:07:47.3543583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3543974Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3544399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3544841Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3545298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 749, in _hash_vectors 2025-09-07T08:07:47.3545808Z random_rotations = torch.randn(rotations_shape, device=vectors.device, dtype=vectors.dtype) 2025-09-07T08:07:47.3546060Z 2025-09-07T08:07:47.3546164Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3546645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3547114Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3547513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3547889Z layer_outputs = layer( 2025-09-07T08:07:47.3548254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3548636Z attn_outputs = self.attention( 2025-09-07T08:07:47.3549032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3549430Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3549826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3550276Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3550735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T08:07:47.3551206Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T08:07:47.3551403Z 2025-09-07T08:07:47.3551514Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3551978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3552443Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3552840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3553225Z layer_outputs = layer( 2025-09-07T08:07:47.3553585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3553970Z attn_outputs = self.attention( 2025-09-07T08:07:47.3554349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3554756Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3555157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3555599Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3556061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T08:07:47.3556519Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T08:07:47.3556707Z 2025-09-07T08:07:47.3556819Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3557294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3557753Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3558179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3558552Z layer_outputs = layer( 2025-09-07T08:07:47.3558909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3559304Z attn_outputs = self.attention( 2025-09-07T08:07:47.3559669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3560079Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3560471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3560907Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3561360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T08:07:47.3561761Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T08:07:47.3561909Z 2025-09-07T08:07:47.3562008Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3562474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3562933Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3563326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3563692Z layer_outputs = layer( 2025-09-07T08:07:47.3564052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3564433Z attn_outputs = self.attention( 2025-09-07T08:07:47.3564804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3565199Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3565580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:07:47.3566065Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:07:47.3566637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:07:47.3567130Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:07:47.3567550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T08:07:47.3568014Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T08:07:47.3568216Z 2025-09-07T08:07:47.3568354Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3568592Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3569070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3569548Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3569947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3570320Z layer_outputs = layer( 2025-09-07T08:07:47.3570685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3571063Z attn_outputs = self.attention( 2025-09-07T08:07:47.3571495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3571893Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3572286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:07:47.3572782Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:07:47.3573367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:07:47.3573846Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:07:47.3574269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T08:07:47.3574683Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T08:07:47.3574831Z 2025-09-07T08:07:47.3574914Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3575112Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3575326Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3575790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3576253Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3576644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3577015Z layer_outputs = layer( 2025-09-07T08:07:47.3577362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3577740Z attn_outputs = self.attention( 2025-09-07T08:07:47.3578115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3578512Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3578899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3579295Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3579696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:07:47.3580196Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3580703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3581160Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3581348Z 2025-09-07T08:07:47.3581425Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3581625Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3581849Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3582314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3582766Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3583159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3583534Z layer_outputs = layer( 2025-09-07T08:07:47.3583892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3584264Z attn_outputs = self.attention( 2025-09-07T08:07:47.3584657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3585053Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3585446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3585860Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3586255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:07:47.3586767Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3587287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:07:47.3587683Z return torch.cat(slices, dim=3) 2025-09-07T08:07:47.3587805Z 2025-09-07T08:07:47.3587886Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3588091Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3588281Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3588501Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3588977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3589440Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3589828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3590205Z layer_outputs = layer( 2025-09-07T08:07:47.3590565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3590946Z attn_outputs = self.attention( 2025-09-07T08:07:47.3591330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3591721Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3592114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3592517Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3592917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:07:47.3593397Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3593884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3594345Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3594541Z 2025-09-07T08:07:47.3594617Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3594843Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3595312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3595767Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3596164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3596540Z layer_outputs = layer( 2025-09-07T08:07:47.3596900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3597282Z attn_outputs = self.attention( 2025-09-07T08:07:47.3597653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3598078Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3598468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3598868Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3599275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:07:47.3599758Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3600247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:07:47.3600642Z return torch.cat(slices, dim=3) 2025-09-07T08:07:47.3600763Z 2025-09-07T08:07:47.3600868Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3601330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3601784Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3602174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3602546Z layer_outputs = layer( 2025-09-07T08:07:47.3602904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3603273Z attn_outputs = self.attention( 2025-09-07T08:07:47.3603642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3604035Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3604422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T08:07:47.3604912Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T08:07:47.3605429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T08:07:47.3605862Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T08:07:47.3606008Z 2025-09-07T08:07:47.3606106Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3606574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3607029Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3607412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3607782Z layer_outputs = layer( 2025-09-07T08:07:47.3608136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3608511Z attn_outputs = self.attention( 2025-09-07T08:07:47.3608883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3609267Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3609656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3610051Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3610447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T08:07:47.3610886Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:07:47.3611076Z 2025-09-07T08:07:47.3611200Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3611672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3612129Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3612561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3612943Z layer_outputs = layer( 2025-09-07T08:07:47.3613292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3613663Z attn_outputs = self.attention( 2025-09-07T08:07:47.3614034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3614425Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3614806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3615209Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3615603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T08:07:47.3616086Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3616583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3617038Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3617227Z 2025-09-07T08:07:47.3617306Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3617533Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3618002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3618461Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3618845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3619220Z layer_outputs = layer( 2025-09-07T08:07:47.3619578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3619952Z attn_outputs = self.attention( 2025-09-07T08:07:47.3620320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3620703Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3621104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3621503Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3621899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:07:47.3622305Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:07:47.3622461Z 2025-09-07T08:07:47.3622558Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3623019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3623474Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3623865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3624232Z layer_outputs = layer( 2025-09-07T08:07:47.3624615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3624991Z attn_outputs = self.attention( 2025-09-07T08:07:47.3625361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3625768Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3626156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3626565Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3626963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:07:47.3627373Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:07:47.3627529Z 2025-09-07T08:07:47.3627615Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3627837Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3628302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3628762Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3629150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3629519Z layer_outputs = layer( 2025-09-07T08:07:47.3629865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3630238Z attn_outputs = self.attention( 2025-09-07T08:07:47.3630610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3631004Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3631395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T08:07:47.3631885Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:07:47.3632424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:07:47.3632908Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:07:47.3633089Z 2025-09-07T08:07:47.3633175Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3633376Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3633591Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3634064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3634519Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3634908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3635280Z layer_outputs = layer( 2025-09-07T08:07:47.3635629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1601, in forward 2025-09-07T08:07:47.3636020Z attn_output = prev_attn_output + attn_output 2025-09-07T08:07:47.3636164Z 2025-09-07T08:07:47.3636237Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3636434Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3636621Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3636815Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3637034Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3637529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3637983Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3638367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3638755Z layer_outputs = layer( 2025-09-07T08:07:47.3639112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:07:47.3639548Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:07:47.3639712Z 2025-09-07T08:07:47.3639795Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3639986Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3640180Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3640374Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3640563Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3640750Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3640971Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3641441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3641894Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3642275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3642648Z layer_outputs = layer( 2025-09-07T08:07:47.3642999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3643372Z attn_outputs = self.attention( 2025-09-07T08:07:47.3643741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3644125Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3644511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:07:47.3644980Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3645469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3645923Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3646108Z 2025-09-07T08:07:47.3646183Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3646378Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3646598Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3647062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3647516Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3647897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3648267Z layer_outputs = layer( 2025-09-07T08:07:47.3648621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3648994Z attn_outputs = self.attention( 2025-09-07T08:07:47.3649355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3649741Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3650134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:07:47.3650636Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3651128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:07:47.3651534Z return torch.cat(slices, dim=3) 2025-09-07T08:07:47.3651663Z 2025-09-07T08:07:47.3651762Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3652247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3652702Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3653093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3653456Z layer_outputs = layer( 2025-09-07T08:07:47.3653811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3654188Z attn_outputs = self.attention( 2025-09-07T08:07:47.3654558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3654955Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3655340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:07:47.3655791Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:07:47.3655988Z 2025-09-07T08:07:47.3656087Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3656551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3657011Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3657398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3657775Z layer_outputs = layer( 2025-09-07T08:07:47.3658133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3658512Z attn_outputs = self.attention( 2025-09-07T08:07:47.3658882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3659265Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3659655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:07:47.3660097Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:07:47.3660283Z 2025-09-07T08:07:47.3660370Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3660565Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3660759Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3660952Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3661169Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3661638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3662084Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3662471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3662838Z layer_outputs = layer( 2025-09-07T08:07:47.3663193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3663592Z attn_outputs = self.attention( 2025-09-07T08:07:47.3663957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3664345Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3664738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T08:07:47.3665237Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3665754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3666211Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3666406Z 2025-09-07T08:07:47.3666481Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3666676Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3666904Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3667362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3667812Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3668216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3668670Z layer_outputs = layer( 2025-09-07T08:07:47.3669041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3669420Z attn_outputs = self.attention( 2025-09-07T08:07:47.3669806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3670202Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3670606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:07:47.3671028Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:07:47.3671187Z 2025-09-07T08:07:47.3671287Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3671750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3672205Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3672600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3672973Z layer_outputs = layer( 2025-09-07T08:07:47.3673321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3673696Z attn_outputs = self.attention( 2025-09-07T08:07:47.3674062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3674449Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3674833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:07:47.3675246Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:07:47.3675409Z 2025-09-07T08:07:47.3675506Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3675967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3676425Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3676860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3677236Z layer_outputs = layer( 2025-09-07T08:07:47.3677594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3677991Z attn_outputs = self.attention( 2025-09-07T08:07:47.3678357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3678773Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3679164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T08:07:47.3679666Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:07:47.3680211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:07:47.3680689Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:07:47.3680869Z 2025-09-07T08:07:47.3680944Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3681146Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3681344Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3681537Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3681725Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3681920Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3682112Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3682305Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3682519Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3682985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3683445Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3683838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3684210Z layer_outputs = layer( 2025-09-07T08:07:47.3684563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3684942Z attn_outputs = self.attention( 2025-09-07T08:07:47.3685312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3685708Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3686104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3686536Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3686992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 749, in _hash_vectors 2025-09-07T08:07:47.3687488Z random_rotations = torch.randn(rotations_shape, device=vectors.device, dtype=vectors.dtype) 2025-09-07T08:07:47.3687717Z 2025-09-07T08:07:47.3687823Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3688298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3688749Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3689142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3689510Z layer_outputs = layer( 2025-09-07T08:07:47.3689895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3690268Z attn_outputs = self.attention( 2025-09-07T08:07:47.3690632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3691027Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3691438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3691893Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3692343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T08:07:47.3692795Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T08:07:47.3692996Z 2025-09-07T08:07:47.3693093Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3693559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3694018Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3694410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3694775Z layer_outputs = layer( 2025-09-07T08:07:47.3695135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3695511Z attn_outputs = self.attention( 2025-09-07T08:07:47.3695877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3696272Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3696655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3697089Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3697534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T08:07:47.3697983Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T08:07:47.3698164Z 2025-09-07T08:07:47.3698271Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3698726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3699176Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3699565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3699936Z layer_outputs = layer( 2025-09-07T08:07:47.3700294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3700666Z attn_outputs = self.attention( 2025-09-07T08:07:47.3701036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3701432Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3701821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:07:47.3702258Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:07:47.3702706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T08:07:47.3703112Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T08:07:47.3703260Z 2025-09-07T08:07:47.3703392Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3703870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3704339Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3704734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3705127Z layer_outputs = layer( 2025-09-07T08:07:47.3705491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3705866Z attn_outputs = self.attention( 2025-09-07T08:07:47.3706231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3706631Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3707023Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:07:47.3707503Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:07:47.3708078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:07:47.3708563Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:07:47.3708982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T08:07:47.3709448Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T08:07:47.3709643Z 2025-09-07T08:07:47.3709730Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3709958Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3710420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3710883Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3711277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3711653Z layer_outputs = layer( 2025-09-07T08:07:47.3712010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3712376Z attn_outputs = self.attention( 2025-09-07T08:07:47.3712747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3713138Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3713532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:07:47.3714016Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:07:47.3714584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:07:47.3715076Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:07:47.3715506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T08:07:47.3715928Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T08:07:47.3716065Z 2025-09-07T08:07:47.3716148Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3716342Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3716600Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3716926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3717010Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3717293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3717382Z layer_outputs = layer( 2025-09-07T08:07:47.3717631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3717706Z attn_outputs = self.attention( 2025-09-07T08:07:47.3717956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3718040Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3718299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3718399Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3718655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:07:47.3718853Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3719126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3719262Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3719265Z 2025-09-07T08:07:47.3719341Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3719422Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3719523Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3719843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3719930Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3720180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3720254Z layer_outputs = layer( 2025-09-07T08:07:47.3720502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3720577Z attn_outputs = self.attention( 2025-09-07T08:07:47.3720823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3720920Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3721168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3721266Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3721519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:07:47.3721710Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3721982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:07:47.3722053Z return torch.cat(slices, dim=3) 2025-09-07T08:07:47.3722057Z 2025-09-07T08:07:47.3722138Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3722211Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3722285Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3722418Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3722733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3722828Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3723098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3723177Z layer_outputs = layer( 2025-09-07T08:07:47.3723443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3723512Z attn_outputs = self.attention( 2025-09-07T08:07:47.3723777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3723862Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3724116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3724221Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3724475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:07:47.3724651Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3724928Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3725074Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3725077Z 2025-09-07T08:07:47.3725150Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3725248Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3725576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3725658Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3725923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3725988Z layer_outputs = layer( 2025-09-07T08:07:47.3726251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3726323Z attn_outputs = self.attention( 2025-09-07T08:07:47.3726581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3726673Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3726930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3727035Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3727290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:07:47.3727457Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3727737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:07:47.3727808Z return torch.cat(slices, dim=3) 2025-09-07T08:07:47.3727811Z 2025-09-07T08:07:47.3727914Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3728238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3728329Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3728619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3728685Z layer_outputs = layer( 2025-09-07T08:07:47.3728946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3729032Z attn_outputs = self.attention( 2025-09-07T08:07:47.3729289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3729391Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3729635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T08:07:47.3729831Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T08:07:47.3730118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T08:07:47.3730214Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T08:07:47.3730217Z 2025-09-07T08:07:47.3730313Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3730631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3730715Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3730963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3731035Z layer_outputs = layer( 2025-09-07T08:07:47.3731283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3731358Z attn_outputs = self.attention( 2025-09-07T08:07:47.3731606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3731696Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3731937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3732032Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3732290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T08:07:47.3732429Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:07:47.3732432Z 2025-09-07T08:07:47.3732533Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3732844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3732934Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3733181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3733247Z layer_outputs = layer( 2025-09-07T08:07:47.3733502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3733572Z attn_outputs = self.attention( 2025-09-07T08:07:47.3733831Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3733912Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3734158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3734287Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3734535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T08:07:47.3734720Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:07:47.3735003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:07:47.3735166Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:07:47.3735169Z 2025-09-07T08:07:47.3735246Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3735343Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3735666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3735751Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3736005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3736068Z layer_outputs = layer( 2025-09-07T08:07:47.3736316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3736396Z attn_outputs = self.attention( 2025-09-07T08:07:47.3736647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3736739Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3736987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3737092Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3737344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:07:47.3737449Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:07:47.3737453Z 2025-09-07T08:07:47.3737557Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3737874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3737969Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3738222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3738295Z layer_outputs = layer( 2025-09-07T08:07:47.3738542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3738614Z attn_outputs = self.attention( 2025-09-07T08:07:47.3738874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3738957Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3739212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:07:47.3739310Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:07:47.3739557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:07:47.3739670Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:07:47.3739673Z 2025-09-07T08:07:47.3739747Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3739849Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3740196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3740287Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3740533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3740613Z layer_outputs = layer( 2025-09-07T08:07:47.3740869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:07:47.3740954Z attn_outputs = self.attention( 2025-09-07T08:07:47.3741213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:07:47.3741296Z self_attention_outputs = self.self_attention( 2025-09-07T08:07:47.3741545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T08:07:47.3741751Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:07:47.3742042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:07:47.3742187Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:07:47.3742190Z 2025-09-07T08:07:47.3742266Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3742348Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3742419Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3742488Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3742566Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3742637Z cudagraph partition due to non gpu ops 2025-09-07T08:07:47.3742741Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3743058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3743142Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3743401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:07:47.3743466Z layer_outputs = layer( 2025-09-07T08:07:47.3743724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:07:47.3743839Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:07:47.3743842Z 2025-09-07T08:07:47.3743945Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:47.3744259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:07:47.3744342Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:07:47.3744599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1753, in forward 2025-09-07T08:07:47.3744704Z return torch.cat([attn_output, hidden_states], dim=-1) 2025-09-07T08:07:47.3744707Z 2025-09-07T08:07:49.8285832Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:49.8286562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2495, in torch_dynamo_resume_in_forward_at_2481 2025-09-07T08:07:49.8287083Z logits = self.lm_head(sequence_output) 2025-09-07T08:07:49.8287493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1886, in forward 2025-09-07T08:07:49.8288025Z return apply_chunking_to_forward(self.forward_chunk, self.chunk_size_lm_head, self.seq_len_dim, hidden_states) 2025-09-07T08:07:49.8288897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:49.8289308Z return forward_fn(*input_tensors) 2025-09-07T08:07:49.8289697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1889, in forward_chunk 2025-09-07T08:07:49.8290170Z hidden_states = self.decoder(hidden_states) 2025-09-07T08:07:49.8290319Z 2025-09-07T08:07:49.8290399Z cudagraph partition due to non gpu ops 2025-09-07T08:07:49.8290663Z cudagraph partition due to non gpu ops 2025-09-07T08:07:50.3050053Z pass 2025-09-07T08:07:50.3050526Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:07:52.0247879Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:07:52.0249442Z import pynvml # type: ignore[import] 2025-09-07T08:07:53.7408293Z 2025-09-07T08:07:58.5373476Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:07:58.5373944Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:07:58.5374321Z cpu eval hf_Roberta_base 2025-09-07T08:07:59.9125894Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:00.1409717Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:00.3673730Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:10.4600052Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4600531Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4600876Z return mod(**inputs) 2025-09-07T08:08:10.4601405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4601912Z outputs = self.roberta( 2025-09-07T08:08:10.4602362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 817, in forward 2025-09-07T08:08:10.4602853Z embedding_output = self.embeddings( 2025-09-07T08:08:10.4603356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 90, in forward 2025-09-07T08:08:10.4604067Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length) 2025-09-07T08:08:10.4604754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1576, in create_position_ids_from_input_ids 2025-09-07T08:08:10.4605300Z mask = input_ids.ne(padding_idx).int() 2025-09-07T08:08:10.4605443Z 2025-09-07T08:08:10.4605562Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4605777Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4606027Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4606296Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4606520Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4606718Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4606921Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4607132Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4607353Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4607556Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4607757Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4607971Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4608216Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4608583Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4609250Z return mod(**inputs) 2025-09-07T08:08:10.4609692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4610200Z outputs = self.roberta( 2025-09-07T08:08:10.4610627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 817, in forward 2025-09-07T08:08:10.4611108Z embedding_output = self.embeddings( 2025-09-07T08:08:10.4611570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 90, in forward 2025-09-07T08:08:10.4612105Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length) 2025-09-07T08:08:10.4612723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1577, in create_position_ids_from_input_ids 2025-09-07T08:08:10.4613307Z incremental_indices = (torch.cumsum(mask, dim=1).type_as(mask) + past_key_values_length) * mask 2025-09-07T08:08:10.4613563Z 2025-09-07T08:08:10.4613667Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4614036Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4614373Z return mod(**inputs) 2025-09-07T08:08:10.4614764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4615252Z outputs = self.roberta( 2025-09-07T08:08:10.4615712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 817, in forward 2025-09-07T08:08:10.4616198Z embedding_output = self.embeddings( 2025-09-07T08:08:10.4616713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 90, in forward 2025-09-07T08:08:10.4617322Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length) 2025-09-07T08:08:10.4618032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1577, in create_position_ids_from_input_ids 2025-09-07T08:08:10.4618598Z incremental_indices = (torch.cumsum(mask, dim=1).type_as(mask) + past_key_values_length) * mask 2025-09-07T08:08:10.4618865Z 2025-09-07T08:08:10.4618977Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4619377Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4619714Z return mod(**inputs) 2025-09-07T08:08:10.4620103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4620552Z outputs = self.roberta( 2025-09-07T08:08:10.4621053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 817, in forward 2025-09-07T08:08:10.4621506Z embedding_output = self.embeddings( 2025-09-07T08:08:10.4621959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 119, in forward 2025-09-07T08:08:10.4622448Z embeddings += position_embeddings 2025-09-07T08:08:10.4622627Z 2025-09-07T08:08:10.4622719Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4622958Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4623223Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4623447Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4623644Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4623870Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4624076Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4624326Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4624742Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4625080Z return mod(**inputs) 2025-09-07T08:08:10.4625450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4625860Z outputs = self.roberta( 2025-09-07T08:08:10.4626232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:08:10.4626647Z encoder_outputs = self.encoder( 2025-09-07T08:08:10.4627035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:08:10.4627410Z layer_outputs = layer_module( 2025-09-07T08:08:10.4627744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:10.4628084Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:10.4628467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:08:10.4628855Z self_attention_outputs = self.attention( 2025-09-07T08:08:10.4629225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4629577Z return func(*args, **kwargs) 2025-09-07T08:08:10.4629941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:08:10.4630317Z self_outputs = self.self( 2025-09-07T08:08:10.4630658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4631008Z return func(*args, **kwargs) 2025-09-07T08:08:10.4631374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:08:10.4631812Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:10.4631993Z 2025-09-07T08:08:10.4632069Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4632270Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4632466Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4632649Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4632848Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4633041Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4633231Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4633414Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4633603Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4633794Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4633985Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4634168Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4634362Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4634581Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4634920Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4635216Z return mod(**inputs) 2025-09-07T08:08:10.4635584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4635964Z outputs = self.roberta( 2025-09-07T08:08:10.4636326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:08:10.4636707Z encoder_outputs = self.encoder( 2025-09-07T08:08:10.4637074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:08:10.4637454Z layer_outputs = layer_module( 2025-09-07T08:08:10.4637817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:10.4638167Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:10.4638552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:08:10.4638978Z self_attention_outputs = self.attention( 2025-09-07T08:08:10.4640673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4641024Z return func(*args, **kwargs) 2025-09-07T08:08:10.4641395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:08:10.4641767Z self_outputs = self.self( 2025-09-07T08:08:10.4642110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4642462Z return func(*args, **kwargs) 2025-09-07T08:08:10.4642828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:08:10.4643264Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:10.4643437Z 2025-09-07T08:08:10.4643510Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4643703Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4643898Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4644091Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4644275Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4644465Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4644659Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4644852Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4645034Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4645234Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4645428Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4645621Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4645807Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4646029Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4646374Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4646677Z return mod(**inputs) 2025-09-07T08:08:10.4647041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4647429Z outputs = self.roberta( 2025-09-07T08:08:10.4647794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:08:10.4648177Z encoder_outputs = self.encoder( 2025-09-07T08:08:10.4648559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:08:10.4648928Z layer_outputs = layer_module( 2025-09-07T08:08:10.4649279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:10.4649624Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:10.4650006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:08:10.4650395Z self_attention_outputs = self.attention( 2025-09-07T08:08:10.4650750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4651101Z return func(*args, **kwargs) 2025-09-07T08:08:10.4651472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:08:10.4651906Z self_outputs = self.self( 2025-09-07T08:08:10.4652245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4652585Z return func(*args, **kwargs) 2025-09-07T08:08:10.4652952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:08:10.4653402Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:10.4653586Z 2025-09-07T08:08:10.4653674Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4653875Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4654066Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4654263Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4654456Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4654652Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4654837Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4655032Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4655224Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4655416Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4655600Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4655790Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4656053Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4656274Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4656604Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4656903Z return mod(**inputs) 2025-09-07T08:08:10.4657263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4657637Z outputs = self.roberta( 2025-09-07T08:08:10.4658003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:08:10.4658372Z encoder_outputs = self.encoder( 2025-09-07T08:08:10.4658746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:08:10.4659123Z layer_outputs = layer_module( 2025-09-07T08:08:10.4659448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:10.4659781Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:10.4660167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:08:10.4660553Z self_attention_outputs = self.attention( 2025-09-07T08:08:10.4660915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4661262Z return func(*args, **kwargs) 2025-09-07T08:08:10.4661626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:08:10.4662003Z self_outputs = self.self( 2025-09-07T08:08:10.4662337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4662686Z return func(*args, **kwargs) 2025-09-07T08:08:10.4663052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:08:10.4663475Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:10.4663651Z 2025-09-07T08:08:10.4663725Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4663920Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4664113Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4664297Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4664490Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4664717Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4664915Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4665100Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4665291Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4665502Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4665697Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4665881Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4666093Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4666324Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4666671Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4666982Z return mod(**inputs) 2025-09-07T08:08:10.4667341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4667729Z outputs = self.roberta( 2025-09-07T08:08:10.4668095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:08:10.4668621Z encoder_outputs = self.encoder( 2025-09-07T08:08:10.4668993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:08:10.4669378Z layer_outputs = layer_module( 2025-09-07T08:08:10.4669705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:10.4670048Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:10.4670431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:08:10.4670812Z self_attention_outputs = self.attention( 2025-09-07T08:08:10.4671183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4671533Z return func(*args, **kwargs) 2025-09-07T08:08:10.4671907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:08:10.4672290Z self_outputs = self.self( 2025-09-07T08:08:10.4672617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4672965Z return func(*args, **kwargs) 2025-09-07T08:08:10.4673334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:08:10.4673764Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:10.4673936Z 2025-09-07T08:08:10.4674012Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4674211Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4674411Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4674609Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4674795Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4674990Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4675182Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4675379Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4675566Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4675761Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4675953Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4676144Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4676335Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4676551Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4676891Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4677196Z return mod(**inputs) 2025-09-07T08:08:10.4677631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4678001Z outputs = self.roberta( 2025-09-07T08:08:10.4678364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:08:10.4678772Z encoder_outputs = self.encoder( 2025-09-07T08:08:10.4679148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:08:10.4679557Z layer_outputs = layer_module( 2025-09-07T08:08:10.4679877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:10.4680222Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:10.4680600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:08:10.4680990Z self_attention_outputs = self.attention( 2025-09-07T08:08:10.4681338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4681683Z return func(*args, **kwargs) 2025-09-07T08:08:10.4682050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:08:10.4682425Z self_outputs = self.self( 2025-09-07T08:08:10.4682758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4683092Z return func(*args, **kwargs) 2025-09-07T08:08:10.4683455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:08:10.4683883Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:10.4684050Z 2025-09-07T08:08:10.4684134Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4684331Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4684514Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4684703Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4684899Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4685090Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4685278Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4685471Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4685664Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4685855Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4686038Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4686230Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4686423Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4686640Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4686974Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4687278Z return mod(**inputs) 2025-09-07T08:08:10.4687642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4688026Z outputs = self.roberta( 2025-09-07T08:08:10.4688383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:08:10.4688764Z encoder_outputs = self.encoder( 2025-09-07T08:08:10.4689139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:08:10.4689514Z layer_outputs = layer_module( 2025-09-07T08:08:10.4689841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:10.4690169Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:10.4690581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:08:10.4690972Z self_attention_outputs = self.attention( 2025-09-07T08:08:10.4691337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4691703Z return func(*args, **kwargs) 2025-09-07T08:08:10.4692087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:08:10.4692467Z self_outputs = self.self( 2025-09-07T08:08:10.4692808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4693155Z return func(*args, **kwargs) 2025-09-07T08:08:10.4693524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:08:10.4693956Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:10.4694134Z 2025-09-07T08:08:10.4694207Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4694407Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4694604Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4694795Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4694992Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4695188Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4695378Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4695560Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4695752Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4695944Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4696137Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4696321Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4696519Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4696737Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4697077Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4697372Z return mod(**inputs) 2025-09-07T08:08:10.4697744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4698126Z outputs = self.roberta( 2025-09-07T08:08:10.4698490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:08:10.4698869Z encoder_outputs = self.encoder( 2025-09-07T08:08:10.4699238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:08:10.4699623Z layer_outputs = layer_module( 2025-09-07T08:08:10.4699959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:10.4700298Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:10.4700685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:08:10.4701071Z self_attention_outputs = self.attention( 2025-09-07T08:08:10.4701435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4701788Z return func(*args, **kwargs) 2025-09-07T08:08:10.4702160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:08:10.4702533Z self_outputs = self.self( 2025-09-07T08:08:10.4702870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4703260Z return func(*args, **kwargs) 2025-09-07T08:08:10.4703635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:08:10.4704063Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:10.4704249Z 2025-09-07T08:08:10.4704321Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4704522Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4704785Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4704978Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4705159Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4705349Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4705538Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4705733Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4705919Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4706109Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4706303Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4706501Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4706686Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4706906Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4707244Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4707547Z return mod(**inputs) 2025-09-07T08:08:10.4707910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4708277Z outputs = self.roberta( 2025-09-07T08:08:10.4708638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:08:10.4709012Z encoder_outputs = self.encoder( 2025-09-07T08:08:10.4709386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:08:10.4709764Z layer_outputs = layer_module( 2025-09-07T08:08:10.4710079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:10.4710413Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:10.4710793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:08:10.4711177Z self_attention_outputs = self.attention( 2025-09-07T08:08:10.4711524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4711868Z return func(*args, **kwargs) 2025-09-07T08:08:10.4712232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:08:10.4712603Z self_outputs = self.self( 2025-09-07T08:08:10.4712940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4713275Z return func(*args, **kwargs) 2025-09-07T08:08:10.4713642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:08:10.4714068Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:10.4714237Z 2025-09-07T08:08:10.4714320Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4714514Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4714702Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4714898Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4715096Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4715291Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4715473Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4715660Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4715881Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4716075Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4716260Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4716456Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4716648Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4716890Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4717218Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4717545Z return mod(**inputs) 2025-09-07T08:08:10.4717907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4718291Z outputs = self.roberta( 2025-09-07T08:08:10.4718649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:08:10.4719037Z encoder_outputs = self.encoder( 2025-09-07T08:08:10.4719413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:08:10.4719792Z layer_outputs = layer_module( 2025-09-07T08:08:10.4720119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:10.4720452Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:10.4720840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:08:10.4721229Z self_attention_outputs = self.attention( 2025-09-07T08:08:10.4721588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4721940Z return func(*args, **kwargs) 2025-09-07T08:08:10.4722308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:08:10.4722689Z self_outputs = self.self( 2025-09-07T08:08:10.4723025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4723375Z return func(*args, **kwargs) 2025-09-07T08:08:10.4723734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:08:10.4724170Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:10.4724345Z 2025-09-07T08:08:10.4724419Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4724618Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4724815Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4724996Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4725185Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4725378Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4725574Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4725762Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4725956Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4726150Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4726344Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4726530Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4726725Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4726948Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4727290Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4727591Z return mod(**inputs) 2025-09-07T08:08:10.4727953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4728335Z outputs = self.roberta( 2025-09-07T08:08:10.4728738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:08:10.4729115Z encoder_outputs = self.encoder( 2025-09-07T08:08:10.4729484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:08:10.4729931Z layer_outputs = layer_module( 2025-09-07T08:08:10.4749645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:10.4750237Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:10.4750655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:08:10.4751056Z self_attention_outputs = self.attention( 2025-09-07T08:08:10.4751435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4751798Z return func(*args, **kwargs) 2025-09-07T08:08:10.4752186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:08:10.4752574Z self_outputs = self.self( 2025-09-07T08:08:10.4752917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4753272Z return func(*args, **kwargs) 2025-09-07T08:08:10.4753648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:08:10.4754087Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:10.4754266Z 2025-09-07T08:08:10.4754348Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4754557Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4754753Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4754957Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4755143Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4755339Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4755537Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4755731Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4755920Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4756112Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4756307Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4756502Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4756694Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4756931Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:10.4757279Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:10.4757593Z return mod(**inputs) 2025-09-07T08:08:10.4757965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:08:10.4758345Z outputs = self.roberta( 2025-09-07T08:08:10.4758714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:08:10.4759102Z encoder_outputs = self.encoder( 2025-09-07T08:08:10.4759483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:08:10.4759871Z layer_outputs = layer_module( 2025-09-07T08:08:10.4760193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:10.4760537Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:10.4760924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:08:10.4761657Z self_attention_outputs = self.attention( 2025-09-07T08:08:10.4762022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4762380Z return func(*args, **kwargs) 2025-09-07T08:08:10.4762757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:08:10.4763164Z self_outputs = self.self( 2025-09-07T08:08:10.4763526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:10.4763871Z return func(*args, **kwargs) 2025-09-07T08:08:10.4764247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:08:10.4764679Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:10.4764846Z 2025-09-07T08:08:10.4764935Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4765138Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4765329Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4765523Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4765720Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4765921Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4766106Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4766304Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4766499Z cudagraph partition due to non gpu ops 2025-09-07T08:08:10.4766693Z cudagraph partition due to non gpu ops 2025-09-07T08:08:24.4182506Z pass 2025-09-07T08:08:24.4182926Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:26.9157407Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:08:26.9158817Z import pynvml # type: ignore[import] 2025-09-07T08:08:28.6316408Z 2025-09-07T08:08:30.8246884Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:08:30.8248005Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:08:30.8248369Z cpu eval hf_T5 2025-09-07T08:08:31.2834007Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:31.5141103Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:31.7458657Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:45.8176106Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:45.8176657Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:45.8177169Z return mod(**inputs) 2025-09-07T08:08:45.8177701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:08:45.8178081Z decoder_outputs = self.decoder( 2025-09-07T08:08:45.8178441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:08:45.8178839Z layer_outputs = layer_module( 2025-09-07T08:08:45.8179327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:45.8179683Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:45.8180252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:08:45.8180781Z self_attention_outputs = self.layer[0]( 2025-09-07T08:08:45.8181286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:08:45.8182130Z attention_output = self.SelfAttention( 2025-09-07T08:08:45.8182494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 546, in forward 2025-09-07T08:08:45.8182858Z position_bias = position_bias + causal_mask 2025-09-07T08:08:45.8183058Z 2025-09-07T08:08:45.8183147Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8183354Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8183604Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8183806Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8184004Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8184217Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8184406Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8184601Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8184797Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8184993Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8185191Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8185393Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8185592Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8185789Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8185979Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8186178Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8186373Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8186570Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8186761Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8186959Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8187156Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8187351Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8187541Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8187749Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8187957Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8188155Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8188348Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8188549Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8188745Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8188944Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8189135Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8189334Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8189532Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8189734Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8189934Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8190123Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8190323Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8190522Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8190721Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8190910Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8191109Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8191307Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8191503Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8191694Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8191894Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8192094Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8192294Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8192486Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8192685Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8192893Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8193090Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8193280Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8193512Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:45.8193864Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:45.8194224Z return mod(**inputs) 2025-09-07T08:08:45.8194557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:08:45.8194912Z encoder_outputs = self.encoder( 2025-09-07T08:08:45.8195254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:08:45.8195619Z layer_outputs = layer_module( 2025-09-07T08:08:45.8195967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:45.8196300Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:45.8196649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:08:45.8197001Z self_attention_outputs = self.layer[0]( 2025-09-07T08:08:45.8197377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:08:45.8197778Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:08:45.8197961Z 2025-09-07T08:08:45.8198038Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8198236Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8198437Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8198633Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8198821Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8199022Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8199221Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8199419Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8199612Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8199868Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8200054Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8200257Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8200448Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8200642Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8200825Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8201015Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8201204Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8201395Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8201574Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8201765Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8201960Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8202152Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8202336Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8202529Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8202726Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8202917Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8203132Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:45.8203478Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:45.8203783Z return mod(**inputs) 2025-09-07T08:08:45.8204116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:08:45.8204466Z decoder_outputs = self.decoder( 2025-09-07T08:08:45.8204801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:08:45.8205147Z layer_outputs = layer_module( 2025-09-07T08:08:45.8205473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:45.8205809Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:45.8206144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:08:45.8206494Z self_attention_outputs = self.layer[0]( 2025-09-07T08:08:45.8207692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:08:45.8208056Z attention_output = self.SelfAttention( 2025-09-07T08:08:45.8208410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:08:45.8208776Z query_states = self.q(hidden_states) 2025-09-07T08:08:45.8208912Z 2025-09-07T08:08:45.8209002Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8209199Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8209392Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8209573Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8209766Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8209959Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8210151Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8210332Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8210525Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8210719Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8210914Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8211097Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8211290Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8211488Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8211707Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:45.8212053Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:45.8212354Z return mod(**inputs) 2025-09-07T08:08:45.8212684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:08:45.8213034Z encoder_outputs = self.encoder( 2025-09-07T08:08:45.8213374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:08:45.8213715Z layer_outputs = layer_module( 2025-09-07T08:08:45.8214042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:45.8214380Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:45.8214727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:08:45.8215080Z self_attention_outputs = self.layer[0]( 2025-09-07T08:08:45.8215425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:08:45.8215820Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:08:45.8215998Z 2025-09-07T08:08:45.8216070Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8216261Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8216445Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8216634Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8216831Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8217021Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8217206Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8217396Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8217592Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8217786Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8217969Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8218168Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8218360Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8218555Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8218750Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8218939Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8219135Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8219326Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8219517Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8219750Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8219944Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8220141Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8220335Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8220521Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8220781Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8220970Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8221182Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8221369Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8221562Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8221759Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8221951Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8222136Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8222329Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8222525Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8222718Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8222905Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8223098Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8223289Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8223480Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8223669Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8223863Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8224055Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8224249Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8224434Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8224624Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8224815Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8225006Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8225190Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8225383Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8225579Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8225800Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:45.8226128Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:45.8226431Z return mod(**inputs) 2025-09-07T08:08:45.8226759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:08:45.8227107Z decoder_outputs = self.decoder( 2025-09-07T08:08:45.8227447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:08:45.8227788Z layer_outputs = layer_module( 2025-09-07T08:08:45.8228111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:45.8228452Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:45.8228803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:08:45.8229146Z self_attention_outputs = self.layer[0]( 2025-09-07T08:08:45.8229493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:08:45.8229887Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:08:45.8230058Z 2025-09-07T08:08:45.8230143Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8230341Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8230525Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8230719Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8230911Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8231125Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8231310Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8231503Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8231731Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8231925Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8232113Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8232306Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8232497Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8232715Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8232906Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8233119Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8233311Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8233506Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8233688Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8233881Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8234073Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8234264Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8234446Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8234642Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8234831Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8235020Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8235203Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8235393Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8235584Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8235777Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8235963Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8236155Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8236346Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8236537Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8236751Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:45.8237088Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:45.8237392Z return mod(**inputs) 2025-09-07T08:08:45.8237723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:08:45.8238069Z decoder_outputs = self.decoder( 2025-09-07T08:08:45.8238399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:08:45.8238747Z layer_outputs = layer_module( 2025-09-07T08:08:45.8239073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:45.8239418Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:45.8239760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:08:45.8240119Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:08:45.8240473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 647, in forward 2025-09-07T08:08:45.8240871Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:08:45.8241040Z 2025-09-07T08:08:45.8241123Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8241311Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8241506Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8241703Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8241894Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8242075Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8242267Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8242460Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8242652Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8242835Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8243027Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8243219Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8243412Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8243596Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8243834Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8244026Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8244217Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8244399Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8244595Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8244803Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8244996Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8245205Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8245394Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8245587Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8245779Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8245967Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8246152Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8246338Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8246560Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:45.8246896Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:45.8247194Z return mod(**inputs) 2025-09-07T08:08:45.8247520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:08:45.8247865Z decoder_outputs = self.decoder( 2025-09-07T08:08:45.8248203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:08:45.8248540Z layer_outputs = layer_module( 2025-09-07T08:08:45.8248864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:45.8249203Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:45.8249553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:08:45.8249916Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:08:45.8250273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 343, in forward 2025-09-07T08:08:45.8250666Z hidden_states = hidden_states + self.dropout(forwarded_states) 2025-09-07T08:08:45.8250841Z 2025-09-07T08:08:45.8250913Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8251108Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8251299Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8251492Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8251685Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8251876Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8252072Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8252256Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8252450Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8252645Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8252837Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8253022Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8253213Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8253402Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8253595Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8253784Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8253977Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8254176Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8254372Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8254575Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8254769Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8254962Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8255156Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8255337Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8255526Z cudagraph partition due to non gpu ops 2025-09-07T08:08:45.8255757Z cudagraph partition due to non gpu ops 2025-09-07T08:09:00.5778910Z pass 2025-09-07T08:09:00.5779334Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:09:02.4924933Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:09:02.4926783Z import pynvml # type: ignore[import] 2025-09-07T08:09:04.2085691Z 2025-09-07T08:09:14.2519814Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:09:14.2520122Z loading model: 0it [00:10, ?it/s] 2025-09-07T08:09:14.2520375Z cpu eval hf_T5_base 2025-09-07T08:09:27.1916935Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:09:35.0912469Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:09:42.9325530Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:10:08.6351227Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6351676Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6352026Z return mod(**inputs) 2025-09-07T08:10:08.6352402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:10:08.6352790Z decoder_outputs = self.decoder( 2025-09-07T08:10:08.6353150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6353501Z layer_outputs = layer_module( 2025-09-07T08:10:08.6353841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6354204Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6354569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:10:08.6354933Z self_attention_outputs = self.layer[0]( 2025-09-07T08:10:08.6355301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:10:08.6355660Z attention_output = self.SelfAttention( 2025-09-07T08:10:08.6356024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 546, in forward 2025-09-07T08:10:08.6356390Z position_bias = position_bias + causal_mask 2025-09-07T08:10:08.6356528Z 2025-09-07T08:10:08.6356618Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6356825Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6357030Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6357232Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6357434Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6357622Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6357818Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6358016Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6358217Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6358415Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6358655Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6358850Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6359049Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6359250Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6359467Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6359662Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6359867Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6360070Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6360274Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6360749Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6360965Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6361169Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6361364Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6361554Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6361814Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6362010Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6362256Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6362449Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6362647Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6362847Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6363047Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6363237Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6363433Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6363629Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6363831Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6364024Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6364225Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6364429Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6364631Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6364824Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6365027Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6365231Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6365427Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6365621Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6365828Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6366024Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6366214Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6366399Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6366591Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6366789Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6366978Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6367175Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6367398Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6367754Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6368068Z return mod(**inputs) 2025-09-07T08:10:08.6368499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:10:08.6368850Z encoder_outputs = self.encoder( 2025-09-07T08:10:08.6369200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6369549Z layer_outputs = layer_module( 2025-09-07T08:10:08.6369885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6370238Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6370597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:10:08.6370962Z self_attention_outputs = self.layer[0]( 2025-09-07T08:10:08.6371332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:10:08.6371745Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:10:08.6371927Z 2025-09-07T08:10:08.6372003Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6372210Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6372408Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6372608Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6372798Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6373000Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6373266Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6373471Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6373677Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6373870Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6374069Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6374304Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6374505Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6374733Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6374931Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6375129Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6375326Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6375520Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6375716Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6375913Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6376109Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6376301Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6376497Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6376698Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6376894Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6377085Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6377288Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6377490Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6377716Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6378066Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6378384Z return mod(**inputs) 2025-09-07T08:10:08.6378728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:10:08.6379092Z encoder_outputs = self.encoder( 2025-09-07T08:10:08.6379442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6379798Z layer_outputs = layer_module( 2025-09-07T08:10:08.6380138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6380491Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6380854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:10:08.6381211Z self_attention_outputs = self.layer[0]( 2025-09-07T08:10:08.6381571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:10:08.6381982Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:10:08.6382162Z 2025-09-07T08:10:08.6382251Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6382455Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6382649Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6382852Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6383051Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6383251Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6383439Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6383639Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6383843Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6384040Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6384231Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6384430Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6384633Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6384829Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6385020Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6385217Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6385415Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6385617Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6385845Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6386046Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6386246Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6386447Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6386635Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6386853Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6387093Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6388007Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6388209Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6388408Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6388627Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6388981Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6389300Z return mod(**inputs) 2025-09-07T08:10:08.6389644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:10:08.6390005Z encoder_outputs = self.encoder( 2025-09-07T08:10:08.6390349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6390707Z layer_outputs = layer_module( 2025-09-07T08:10:08.6391039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6391391Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6391740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:10:08.6392103Z self_attention_outputs = self.layer[0]( 2025-09-07T08:10:08.6392463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:10:08.6392868Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:10:08.6393053Z 2025-09-07T08:10:08.6393138Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6393335Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6393539Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6393738Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6393938Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6394128Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6394335Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6394533Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6394729Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6394918Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6395118Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6395321Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6395520Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6395713Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6395911Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6396112Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6396311Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6396500Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6396696Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6396896Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6397095Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6397287Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6397483Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6397681Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6397876Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6398075Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6398263Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6398461Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6398685Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6399102Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6399413Z return mod(**inputs) 2025-09-07T08:10:08.6399745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:10:08.6400128Z encoder_outputs = self.encoder( 2025-09-07T08:10:08.6400483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6400846Z layer_outputs = layer_module( 2025-09-07T08:10:08.6401179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6401522Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6401882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:10:08.6402243Z self_attention_outputs = self.layer[0]( 2025-09-07T08:10:08.6402602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:10:08.6403007Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:10:08.6403193Z 2025-09-07T08:10:08.6403269Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6403473Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6403663Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6403864Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6404065Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6404263Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6404465Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6404658Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6404866Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6405063Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6405264Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6405456Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6405654Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6405852Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6406049Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6406244Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6406447Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6406647Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6406850Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6407041Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6407242Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6407441Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6407640Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6407834Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6408036Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6408235Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6408468Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6408810Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6409122Z return mod(**inputs) 2025-09-07T08:10:08.6409455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:10:08.6409811Z decoder_outputs = self.decoder( 2025-09-07T08:10:08.6410165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6410516Z layer_outputs = layer_module( 2025-09-07T08:10:08.6410849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6411198Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6411586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:10:08.6411944Z self_attention_outputs = self.layer[0]( 2025-09-07T08:10:08.6412306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:10:08.6412673Z attention_output = self.SelfAttention( 2025-09-07T08:10:08.6413058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:10:08.6413441Z query_states = self.q(hidden_states) 2025-09-07T08:10:08.6413571Z 2025-09-07T08:10:08.6413649Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6413849Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6414055Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6414256Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6414447Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6414648Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6414853Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6415054Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6415251Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6415452Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6415655Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6415857Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6416045Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6416246Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6416474Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6416827Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6417138Z return mod(**inputs) 2025-09-07T08:10:08.6417474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:10:08.6417838Z encoder_outputs = self.encoder( 2025-09-07T08:10:08.6418194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6418554Z layer_outputs = layer_module( 2025-09-07T08:10:08.6418885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6419240Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6419606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:10:08.6419967Z self_attention_outputs = self.layer[0]( 2025-09-07T08:10:08.6420319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:10:08.6420725Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:10:08.6420910Z 2025-09-07T08:10:08.6420985Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6421188Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6421390Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6421580Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6421777Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6421974Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6422172Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6422365Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6422569Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6422768Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6422966Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6423157Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6423355Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6423555Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6423751Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6423942Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6424138Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6424370Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6424568Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6424761Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6424960Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6425158Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6425370Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6425561Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6425796Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6425999Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6426203Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6426395Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6426600Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6426800Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6427001Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6427192Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6427399Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6427598Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6427800Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6427989Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6428189Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6428392Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6428589Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6428793Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6428986Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6429183Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6429394Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6429592Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6429782Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6429981Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6430179Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6430376Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6430566Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6430763Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6430991Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6431337Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6431644Z return mod(**inputs) 2025-09-07T08:10:08.6431983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:10:08.6432341Z decoder_outputs = self.decoder( 2025-09-07T08:10:08.6432690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6433039Z layer_outputs = layer_module( 2025-09-07T08:10:08.6433372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6433719Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6434074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:10:08.6434433Z self_attention_outputs = self.layer[0]( 2025-09-07T08:10:08.6434785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:10:08.6435191Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:10:08.6435376Z 2025-09-07T08:10:08.6435451Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6435652Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6435851Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6436038Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6436235Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6436437Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6436668Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6436857Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6437054Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6437251Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6437451Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6437659Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6437861Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6438058Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6438277Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6438469Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6438669Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6438869Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6439135Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6439325Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6439522Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6439723Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6439919Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6440107Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6440306Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6440503Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6440700Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6440890Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6441087Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6441286Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6441484Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6441675Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6441870Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6442065Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6442290Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6442625Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6442936Z return mod(**inputs) 2025-09-07T08:10:08.6443267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:10:08.6443623Z decoder_outputs = self.decoder( 2025-09-07T08:10:08.6443972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6444318Z layer_outputs = layer_module( 2025-09-07T08:10:08.6444651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6444994Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6445350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:10:08.6445711Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:10:08.6446065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 647, in forward 2025-09-07T08:10:08.6446469Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:10:08.6446647Z 2025-09-07T08:10:08.6446725Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6446922Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6447121Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6447317Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6447517Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6447711Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6447899Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6448096Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6448295Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6448491Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6448678Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6448877Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6449134Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6449338Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6449530Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6449731Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6449932Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6450151Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6450341Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6450561Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6450759Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6450962Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6451152Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6451350Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6451545Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6451744Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6451942Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6452139Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6452368Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6452712Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6453024Z return mod(**inputs) 2025-09-07T08:10:08.6453350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:10:08.6453706Z decoder_outputs = self.decoder( 2025-09-07T08:10:08.6454053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6454408Z layer_outputs = layer_module( 2025-09-07T08:10:08.6454735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6455076Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6455434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:10:08.6455808Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:10:08.6456180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 343, in forward 2025-09-07T08:10:08.6456578Z hidden_states = hidden_states + self.dropout(forwarded_states) 2025-09-07T08:10:08.6456752Z 2025-09-07T08:10:08.6456826Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6457020Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6457210Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6457396Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6457586Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6457775Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6457967Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6458150Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6458343Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6458538Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6458728Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6458921Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6459106Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6459300Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6459490Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6459681Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6459869Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6460061Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6460252Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6460449Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6460636Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6460826Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6461020Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6461247Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6461438Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6461634Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6461832Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6462021Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6462226Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6462426Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6462626Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6462836Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6463019Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6463208Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6463432Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6463770Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6464065Z return mod(**inputs) 2025-09-07T08:10:08.6464396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:10:08.6464744Z decoder_outputs = self.decoder( 2025-09-07T08:10:08.6465086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6465432Z layer_outputs = layer_module( 2025-09-07T08:10:08.6465748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6466088Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6466434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:10:08.6466783Z self_attention_outputs = self.layer[0]( 2025-09-07T08:10:08.6467121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:10:08.6467512Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:10:08.6467690Z 2025-09-07T08:10:08.6467762Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6467958Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6468150Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6468384Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6468581Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6468774Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6468975Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6469163Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6469360Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6469557Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6469752Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6469940Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6470137Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6470335Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6470534Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6470723Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6470914Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6471108Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6471302Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6471488Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6471685Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6471879Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6472071Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6472260Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6472452Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6472641Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6472839Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6473027Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6473218Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6473477Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6473673Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6473860Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6474052Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6474243Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6474496Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6474827Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6475155Z return mod(**inputs) 2025-09-07T08:10:08.6475482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:10:08.6475829Z decoder_outputs = self.decoder( 2025-09-07T08:10:08.6476169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6476507Z layer_outputs = layer_module( 2025-09-07T08:10:08.6476839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6477176Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6477521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:10:08.6477868Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:10:08.6478217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 647, in forward 2025-09-07T08:10:08.6478612Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:10:08.6478777Z 2025-09-07T08:10:08.6478861Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6479051Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6479234Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6479428Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6479622Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6479811Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6479997Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6480192Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6480382Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6480576Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6480762Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6480958Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6481154Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6481347Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6481532Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6481723Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6481914Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6482112Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6482296Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6482488Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6482684Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6482875Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6483057Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6483249Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6483445Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6483640Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6483827Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6484023Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6484247Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6484585Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6484884Z return mod(**inputs) 2025-09-07T08:10:08.6485201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:10:08.6485548Z decoder_outputs = self.decoder( 2025-09-07T08:10:08.6486248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6486601Z layer_outputs = layer_module( 2025-09-07T08:10:08.6486920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6487274Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6487658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:10:08.6488030Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:10:08.6488395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 343, in forward 2025-09-07T08:10:08.6488785Z hidden_states = hidden_states + self.dropout(forwarded_states) 2025-09-07T08:10:08.6488961Z 2025-09-07T08:10:08.6489035Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6489233Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6489428Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6489611Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6489804Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6489995Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6490189Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6490373Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6490568Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6490758Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6490953Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6491137Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6491331Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6491524Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6491731Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6491924Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6492115Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6492309Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6492506Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6492698Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6492882Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6493085Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6493281Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6493477Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6493664Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6493857Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6494047Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6494240Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6494424Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6494614Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6494806Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6495001Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6495187Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6495381Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6495603Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6495942Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6496241Z return mod(**inputs) 2025-09-07T08:10:08.6496570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:10:08.6496919Z decoder_outputs = self.decoder( 2025-09-07T08:10:08.6497259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6497605Z layer_outputs = layer_module( 2025-09-07T08:10:08.6497926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6498302Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6498655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:10:08.6499008Z self_attention_outputs = self.layer[0]( 2025-09-07T08:10:08.6499363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:10:08.6499757Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:10:08.6499950Z 2025-09-07T08:10:08.6500021Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6500216Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6500409Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6500593Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6500783Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6500980Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6501170Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6501354Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6501546Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6501740Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6501933Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6502117Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6502309Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6502505Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6502697Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6502884Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6503076Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6503267Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6503457Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6503638Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6503834Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6504029Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6504221Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6504402Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6504592Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6504785Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6504976Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6505157Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6505347Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6505543Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6505736Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6505916Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6506108Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6506297Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6506521Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:08.6506851Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:10:08.6507154Z return mod(**inputs) 2025-09-07T08:10:08.6507477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:10:08.6507829Z decoder_outputs = self.decoder( 2025-09-07T08:10:08.6508170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:10:08.6508504Z layer_outputs = layer_module( 2025-09-07T08:10:08.6508830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:08.6509166Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:08.6509509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:10:08.6509851Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:10:08.6510223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 647, in forward 2025-09-07T08:10:08.6510619Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:10:08.6510787Z 2025-09-07T08:10:08.6510867Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6511062Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6511264Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6511455Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6511693Z cudagraph partition due to non gpu ops 2025-09-07T08:10:08.6511887Z cudagraph partition due to non gpu ops 2025-09-07T08:10:43.5585260Z pass 2025-09-07T08:10:43.5585686Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:10:46.8038489Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:10:46.8039915Z import pynvml # type: ignore[import] 2025-09-07T08:10:48.5218287Z 2025-09-07T08:10:58.0892670Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:10:58.0892982Z loading model: 0it [00:09, ?it/s] 2025-09-07T08:10:58.0893246Z cpu eval hf_T5_large 2025-09-07T08:10:58.4782543Z pass_due_to_skip 2025-09-07T08:10:58.4782941Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:11:00.0029076Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:11:00.0030469Z import pynvml # type: ignore[import] 2025-09-07T08:11:01.7186089Z 2025-09-07T08:11:06.0115638Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:11:06.0116146Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:11:06.0118906Z cpu eval hf_distil_whisper 2025-09-07T08:11:10.3161089Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:11:11.0628078Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:11:11.8083350Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:11:29.3530362Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3530796Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3531126Z return mod(*inputs) 2025-09-07T08:11:29.3531514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3531911Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3532331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 677, in forward 2025-09-07T08:11:29.3532764Z inputs_embeds = nn.functional.gelu(self.conv1(input_features)) 2025-09-07T08:11:29.3532937Z 2025-09-07T08:11:29.3533043Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3533401Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3533720Z return mod(*inputs) 2025-09-07T08:11:29.3534078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1615, in forward 2025-09-07T08:11:29.3534464Z logits = self.classifier(pooled_output) 2025-09-07T08:11:29.3534596Z 2025-09-07T08:11:29.3534692Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3535027Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3535562Z return mod(*inputs) 2025-09-07T08:11:29.3535946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3536323Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3536695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 677, in forward 2025-09-07T08:11:29.3537162Z inputs_embeds = nn.functional.gelu(self.conv1(input_features)) 2025-09-07T08:11:29.3537382Z 2025-09-07T08:11:29.3537481Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3537816Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3538117Z return mod(*inputs) 2025-09-07T08:11:29.3538461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3538836Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3539206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 677, in forward 2025-09-07T08:11:29.3539620Z inputs_embeds = nn.functional.gelu(self.conv1(input_features)) 2025-09-07T08:11:29.3539785Z 2025-09-07T08:11:29.3539883Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3540265Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3540578Z return mod(*inputs) 2025-09-07T08:11:29.3540939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3541325Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3541695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 678, in forward 2025-09-07T08:11:29.3542128Z inputs_embeds = nn.functional.gelu(self.conv2(inputs_embeds)) 2025-09-07T08:11:29.3542307Z 2025-09-07T08:11:29.3542408Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3542754Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3543062Z return mod(*inputs) 2025-09-07T08:11:29.3543409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3543797Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3544176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3544559Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3544931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3545289Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3545679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 409, in forward 2025-09-07T08:11:29.3546100Z hidden_states = self.self_attn_layer_norm(hidden_states) 2025-09-07T08:11:29.3546256Z 2025-09-07T08:11:29.3546362Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3546705Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3547010Z return mod(*inputs) 2025-09-07T08:11:29.3547363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3547751Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3548127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3548502Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3548890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3549250Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3549732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3550136Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3550560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:11:29.3551005Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:11:29.3551172Z 2025-09-07T08:11:29.3551253Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3551468Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3551672Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3551868Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3552067Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3552268Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3552469Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3552657Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3552854Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3553048Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3553275Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3553616Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3553923Z return mod(*inputs) 2025-09-07T08:11:29.3554282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3554663Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3555029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3555413Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3555759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3556112Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3556500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3556897Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3557295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3557706Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3558157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3558643Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3558825Z 2025-09-07T08:11:29.3558930Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3559287Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3559609Z return mod(*inputs) 2025-09-07T08:11:29.3559976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3560370Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3560745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3561135Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3561481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3561843Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3562218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3562650Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3563048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3563465Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3563907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3564364Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3564533Z 2025-09-07T08:11:29.3564610Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3564817Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3565050Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3565399Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3565703Z return mod(*inputs) 2025-09-07T08:11:29.3566062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3566442Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3566816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3567187Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3567528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3567878Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3568349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:11:29.3568868Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:11:29.3569033Z 2025-09-07T08:11:29.3569135Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3569373Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3569614Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3569867Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3570175Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3570621Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3571012Z return mod(*inputs) 2025-09-07T08:11:29.3571444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3571890Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3572294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3572713Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3573098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3573484Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3573922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3574403Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3574962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:11:29.3575418Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:11:29.3575638Z 2025-09-07T08:11:29.3575724Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3575952Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3576153Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3576421Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3576620Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3576855Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3577201Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3577436Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3577674Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3577939Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3578215Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3578691Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3579081Z return mod(*inputs) 2025-09-07T08:11:29.3579539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3580006Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3580515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3581007Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3581451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3581853Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3582365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3582768Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3583249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3583736Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3584281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3584799Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3584991Z 2025-09-07T08:11:29.3585125Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3585508Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3585844Z return mod(*inputs) 2025-09-07T08:11:29.3586300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3586817Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3587289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3587773Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3588181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3588598Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3589102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3589538Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3589972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3590401Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3590894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3591414Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3591604Z 2025-09-07T08:11:29.3591697Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3591974Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3592250Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3592701Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3593087Z return mod(*inputs) 2025-09-07T08:11:29.3593608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3594097Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3594559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3595816Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3596195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3596645Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3597061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:11:29.3597542Z hidden_states = residual + hidden_states 2025-09-07T08:11:29.3597723Z 2025-09-07T08:11:29.3597803Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3598035Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3598299Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3598513Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3598739Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3598948Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3599190Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3599400Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3599621Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3599851Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3600047Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3600269Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3600495Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3600731Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3600972Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3601375Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3601759Z return mod(*inputs) 2025-09-07T08:11:29.3602158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3602609Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3602981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3603435Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3603894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3604329Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3604819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3605330Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3605773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3606216Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3606683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3607197Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3607394Z 2025-09-07T08:11:29.3607505Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3607857Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3608172Z return mod(*inputs) 2025-09-07T08:11:29.3608530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3608904Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3609331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3609709Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3610046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3610395Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3610788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3611209Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3611607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3612010Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3612436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3612873Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3613034Z 2025-09-07T08:11:29.3613113Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3613321Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3613524Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3613718Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3613915Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3614114Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3614314Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3614502Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3614696Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3614892Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3615089Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3615278Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3615472Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3615670Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3615865Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3616053Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3616281Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3616628Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3616942Z return mod(*inputs) 2025-09-07T08:11:29.3617302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3617681Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3618057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3618443Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3618783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3619137Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3619520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3619922Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3620323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3620737Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3621156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3621612Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3621794Z 2025-09-07T08:11:29.3621894Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3622241Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3622581Z return mod(*inputs) 2025-09-07T08:11:29.3622933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3623318Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3623714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3624109Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3624448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3624791Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3625178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3625577Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3625979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3626378Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3626804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3627245Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3627411Z 2025-09-07T08:11:29.3627488Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3627699Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3627926Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3628272Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3628583Z return mod(*inputs) 2025-09-07T08:11:29.3628943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3629337Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3629708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3630093Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3630435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3630783Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3631165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:11:29.3631556Z hidden_states = residual + hidden_states 2025-09-07T08:11:29.3631691Z 2025-09-07T08:11:29.3631765Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3631969Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3632167Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3632371Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3632568Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3632771Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3632968Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3633162Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3633364Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3633562Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3633761Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3633958Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3634156Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3634355Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3634584Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3634924Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3635239Z return mod(*inputs) 2025-09-07T08:11:29.3635647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3636033Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3636399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3636794Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3637132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3637500Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3637880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3638271Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3638667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3639071Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3639493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3639949Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3640125Z 2025-09-07T08:11:29.3640225Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3640572Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3640884Z return mod(*inputs) 2025-09-07T08:11:29.3641241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3641620Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3641985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3642364Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3642700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3643048Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3643419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3643816Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3644206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3644604Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3645020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3645445Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3645603Z 2025-09-07T08:11:29.3645682Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3645887Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3646086Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3646278Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3646477Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3646674Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3646865Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3647063Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3647249Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3647444Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3647641Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3647835Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3648026Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3648222Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3648418Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3648642Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3648862Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3649208Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3649524Z return mod(*inputs) 2025-09-07T08:11:29.3649902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3650312Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3650685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3651064Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3651402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3651748Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3652124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3652523Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3652916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3653319Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3653740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3654194Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3654374Z 2025-09-07T08:11:29.3654474Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3654815Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3655125Z return mod(*inputs) 2025-09-07T08:11:29.3655474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3655857Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3656230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3656612Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3656947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3657291Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3657676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3658073Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3658470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3658874Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3659287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3659721Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3659882Z 2025-09-07T08:11:29.3659958Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3660164Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3660385Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3660732Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3661039Z return mod(*inputs) 2025-09-07T08:11:29.3661391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3661778Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3662164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3662533Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3662861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3663216Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3663578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:11:29.3663975Z hidden_states = residual + hidden_states 2025-09-07T08:11:29.3664109Z 2025-09-07T08:11:29.3664184Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3664382Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3664580Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3664768Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3664959Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3665161Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3665354Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3665540Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3665733Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3665928Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3666122Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3666309Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3666512Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3666704Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3666925Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3667251Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3667550Z return mod(*inputs) 2025-09-07T08:11:29.3667901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3668334Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3668796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3669173Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3669511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3669886Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3670261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3670646Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3671029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3671424Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3671839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3672286Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3672453Z 2025-09-07T08:11:29.3672550Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3672889Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3673192Z return mod(*inputs) 2025-09-07T08:11:29.3673535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3673909Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3674266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3674632Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3675020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3675357Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3675718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3676128Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3676514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3676932Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3677352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3677776Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3677938Z 2025-09-07T08:11:29.3678015Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3678225Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3678429Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3678631Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3678825Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3679024Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3679227Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3679429Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3679627Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3679825Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3680029Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3680227Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3680419Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3680616Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3680814Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3681014Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3681232Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3681569Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3681876Z return mod(*inputs) 2025-09-07T08:11:29.3682233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3682608Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3682980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3683351Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3683691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3684034Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3684405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3684799Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3685188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3685584Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3686005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3686445Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3686621Z 2025-09-07T08:11:29.3686722Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3687063Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3687367Z return mod(*inputs) 2025-09-07T08:11:29.3687751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3688130Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3688494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3688866Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3689205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3689550Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3689919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3690300Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3690686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3691076Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3691481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3691906Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3692061Z 2025-09-07T08:11:29.3692140Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3692338Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3692555Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3692891Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3693196Z return mod(*inputs) 2025-09-07T08:11:29.3693546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3693920Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3694283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3694649Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3694975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3695314Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3695686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:11:29.3696060Z hidden_states = residual + hidden_states 2025-09-07T08:11:29.3696195Z 2025-09-07T08:11:29.3696268Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3696468Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3696665Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3696852Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3697050Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3697240Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3697434Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3697620Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3697817Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3698012Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3698204Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3698391Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3698586Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3698780Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3699003Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3699329Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3699635Z return mod(*inputs) 2025-09-07T08:11:29.3699985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3700355Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3701149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3701521Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3701853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3702201Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3702575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3702983Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3703360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3703753Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3704169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3704613Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3704778Z 2025-09-07T08:11:29.3704874Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3705210Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3705516Z return mod(*inputs) 2025-09-07T08:11:29.3705871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3706242Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3706599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3706970Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3707301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3707642Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3708018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3708398Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3708787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3709179Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3709590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3710017Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3710165Z 2025-09-07T08:11:29.3710240Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3710441Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3710641Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3710836Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3711021Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3711216Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3711536Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3711735Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3711921Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3712117Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3712312Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3712507Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3712693Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3712894Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3713091Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3713282Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3713493Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3713855Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3714163Z return mod(*inputs) 2025-09-07T08:11:29.3714516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3714904Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3715259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3715640Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3715970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3716308Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3716672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3717062Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3717445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3717835Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3718246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3718679Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3718858Z 2025-09-07T08:11:29.3718956Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3719295Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3719603Z return mod(*inputs) 2025-09-07T08:11:29.3719953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3720326Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3720694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3721061Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3721391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3721731Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3722098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3722486Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3722871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3723259Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3723665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3724092Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3724251Z 2025-09-07T08:11:29.3724327Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3724537Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3724776Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3725111Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3725426Z return mod(*inputs) 2025-09-07T08:11:29.3725781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3726165Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3726534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3726949Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3727291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3727640Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3728024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:11:29.3728420Z hidden_states = residual + hidden_states 2025-09-07T08:11:29.3728586Z 2025-09-07T08:11:29.3728662Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3728877Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3729087Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3729283Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3729488Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3729693Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3729897Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3730094Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3730300Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3730503Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3730705Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3730900Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3731108Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3731307Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3731539Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3731890Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3732199Z return mod(*inputs) 2025-09-07T08:11:29.3732562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3732952Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3733337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3733717Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3734065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3734423Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3734816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3735222Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3735617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3736027Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3736459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3736927Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3737103Z 2025-09-07T08:11:29.3737219Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3737565Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3737885Z return mod(*inputs) 2025-09-07T08:11:29.3738251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3738642Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3739015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3739402Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3739745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3740096Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3740523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3740914Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3741306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3741723Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3742164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3742598Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3742751Z 2025-09-07T08:11:29.3742827Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3743033Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3743235Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3743437Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3743633Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3743833Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3744028Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3744224Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3744413Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3744611Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3744804Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3744999Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3745186Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3745385Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3745581Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3745776Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3745995Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3746334Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3746650Z return mod(*inputs) 2025-09-07T08:11:29.3747010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3747388Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3747759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3748140Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3748470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3748814Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3749198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3749587Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3749979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3750378Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3750800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3751247Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3751430Z 2025-09-07T08:11:29.3751528Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3751869Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3752174Z return mod(*inputs) 2025-09-07T08:11:29.3752534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3752907Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3753310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3753686Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3754020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3754381Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3754759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3755170Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3755567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3755966Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3756373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3756811Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3756972Z 2025-09-07T08:11:29.3757048Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3757249Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3757477Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3757811Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3758126Z return mod(*inputs) 2025-09-07T08:11:29.3758481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3758858Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3759227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3759594Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3759928Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3760274Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3760657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:11:29.3761034Z hidden_states = residual + hidden_states 2025-09-07T08:11:29.3761170Z 2025-09-07T08:11:29.3761244Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3761449Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3761649Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3761850Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3762039Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3762236Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3762435Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3762630Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3762818Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3763016Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3763212Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3763410Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3763599Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3763802Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3764027Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3764372Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3764680Z return mod(*inputs) 2025-09-07T08:11:29.3765040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3765421Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3765794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3766198Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3766536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3766877Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3767260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3767676Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3768073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3768555Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3769060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3769526Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3769697Z 2025-09-07T08:11:29.3769810Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3770153Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3770469Z return mod(*inputs) 2025-09-07T08:11:29.3770830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3771216Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3771593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3771966Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3772305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3772653Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3773041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3773450Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3773836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3774226Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3774636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3775062Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3775212Z 2025-09-07T08:11:29.3775287Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3775486Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3775680Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3775871Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3776054Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3776250Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3776441Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3776633Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3776821Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3777015Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3777214Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3777408Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3777591Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3777783Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3777973Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3778171Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3778394Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3778720Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3779023Z return mod(*inputs) 2025-09-07T08:11:29.3779429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3779808Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3780170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3780584Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3780915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3781287Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3781672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3782053Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3782443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3782844Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3783264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3783713Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3783886Z 2025-09-07T08:11:29.3783989Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3784330Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3784641Z return mod(*inputs) 2025-09-07T08:11:29.3784995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3785369Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3785744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3786124Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3786457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3786801Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3787174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3787569Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3787960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3788352Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3788767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3789192Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3789354Z 2025-09-07T08:11:29.3789430Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3789635Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3789867Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3790198Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3790510Z return mod(*inputs) 2025-09-07T08:11:29.3790862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3791242Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3791612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3791981Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3792313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3792687Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3793064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:11:29.3793433Z hidden_states = residual + hidden_states 2025-09-07T08:11:29.3793588Z 2025-09-07T08:11:29.3793660Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3793859Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3794072Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3794267Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3794454Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3794645Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3794838Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3795028Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3795213Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3795406Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3795600Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3795791Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3795978Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3796166Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3796387Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3796722Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3797021Z return mod(*inputs) 2025-09-07T08:11:29.3797370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3797743Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3798112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3798478Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3798802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3799140Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3799513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3799902Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3800282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3800678Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3801093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3801542Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3801711Z 2025-09-07T08:11:29.3801814Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3802145Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3802451Z return mod(*inputs) 2025-09-07T08:11:29.3802801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3803180Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3803547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3803906Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3804233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3804565Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3804936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3805353Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3805732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3806121Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3806548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3806987Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3807138Z 2025-09-07T08:11:29.3807212Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3807411Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3807607Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3807803Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3807994Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3808177Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3808376Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3808568Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3808758Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3808947Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3809136Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3809330Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3809519Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3809706Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3809901Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3810095Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3810317Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3810644Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3810946Z return mod(*inputs) 2025-09-07T08:11:29.3811302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3811677Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3812042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3812406Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3812737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3813074Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3813450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3813830Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3814215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3814606Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3815022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3815469Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3815637Z 2025-09-07T08:11:29.3815738Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3816074Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3816376Z return mod(*inputs) 2025-09-07T08:11:29.3816724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3817097Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3817453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3817822Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3818183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3818527Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3818893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3819292Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3819673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3820079Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3820487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3820903Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3821058Z 2025-09-07T08:11:29.3821132Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3821335Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3821558Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3821880Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3822184Z return mod(*inputs) 2025-09-07T08:11:29.3822532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3822910Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3823271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3823628Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3823955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3824295Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3824670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:11:29.3825045Z hidden_states = residual + hidden_states 2025-09-07T08:11:29.3825172Z 2025-09-07T08:11:29.3825244Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3825442Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3825635Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3825830Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3826019Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3826214Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3826406Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3826595Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3826778Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3826971Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3827164Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3827358Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3827546Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3827737Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3827959Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3828293Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3828589Z return mod(*inputs) 2025-09-07T08:11:29.3828939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3829314Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3829680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3830049Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3830368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3830750Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3831123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3831508Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3831905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3832310Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3832728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3833173Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3833342Z 2025-09-07T08:11:29.3833448Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3833783Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3834091Z return mod(*inputs) 2025-09-07T08:11:29.3834442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3834822Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3835191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3835563Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3835893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3836232Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3836610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3837004Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3837386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3837785Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3838206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3838638Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3838791Z 2025-09-07T08:11:29.3838873Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3839070Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3839267Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3839471Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3839667Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3839858Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3840057Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3840252Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3840455Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3840639Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3840836Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3841033Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3841233Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3841422Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3841616Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3841810Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3842030Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3842359Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3842664Z return mod(*inputs) 2025-09-07T08:11:29.3843018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3843423Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3843791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3844156Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3844483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3844838Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3845231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3845612Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3845995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3846383Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3846804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3847248Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3847416Z 2025-09-07T08:11:29.3847513Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3847854Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3848154Z return mod(*inputs) 2025-09-07T08:11:29.3848502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3848871Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3849230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3849595Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3849928Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3850267Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3850634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3851018Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3851405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3851799Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3852208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3852623Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3852780Z 2025-09-07T08:11:29.3852854Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3853054Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3853279Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3853615Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3853911Z return mod(*inputs) 2025-09-07T08:11:29.3854253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3854629Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3854994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3855352Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3855680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3856016Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3856417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:11:29.3856799Z hidden_states = residual + hidden_states 2025-09-07T08:11:29.3856929Z 2025-09-07T08:11:29.3857003Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3857203Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3857420Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3857615Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3857800Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3858018Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3858209Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3858403Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3858586Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3858781Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3858974Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3859167Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3859348Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3859540Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3859763Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3860097Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3860396Z return mod(*inputs) 2025-09-07T08:11:29.3860743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3861115Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3861482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3861851Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3862057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3862132Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3862385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3862471Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3862722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3862813Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3863089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3863208Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3863211Z 2025-09-07T08:11:29.3863309Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3863500Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3863563Z return mod(*inputs) 2025-09-07T08:11:29.3863818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3863887Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3864132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3864208Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3864413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3864496Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3864737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3864830Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3865106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3865199Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3865475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3865593Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3865596Z 2025-09-07T08:11:29.3865680Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3865768Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3865842Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3865919Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3865989Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3866068Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3866140Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3866212Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3866292Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3866366Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3866443Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3866513Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3866581Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3866660Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3866731Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3866808Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3866905Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3867088Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3867156Z return mod(*inputs) 2025-09-07T08:11:29.3867405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3867479Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3867724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3867793Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3868010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3868085Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3868385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3868476Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3868718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3868817Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3869086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3869214Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3869217Z 2025-09-07T08:11:29.3869315Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3869507Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3869573Z return mod(*inputs) 2025-09-07T08:11:29.3869820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3869900Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3870145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3870225Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3870433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3870568Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3870820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3870908Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3871158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3871267Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3871564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3871666Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3871669Z 2025-09-07T08:11:29.3871744Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3871825Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3871925Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3872116Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3872179Z return mod(*inputs) 2025-09-07T08:11:29.3872424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3872503Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3872745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3872823Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3873026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3873100Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3873352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:11:29.3873427Z hidden_states = residual + hidden_states 2025-09-07T08:11:29.3873430Z 2025-09-07T08:11:29.3873509Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3873581Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3873659Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3873730Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3873801Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3873876Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3873948Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3874017Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3874093Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3874164Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3874243Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3874313Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3874383Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3874462Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3874560Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3874752Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3874814Z return mod(*inputs) 2025-09-07T08:11:29.3875056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3875133Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3875377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3875452Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3875655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3875736Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3876002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3876087Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3876338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3876441Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3876714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3876854Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3876857Z 2025-09-07T08:11:29.3876952Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3877140Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3877202Z return mod(*inputs) 2025-09-07T08:11:29.3877458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3877526Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3877773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3877842Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3878047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3878129Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3878368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3878462Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3878704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3878793Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3879064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3879165Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3879170Z 2025-09-07T08:11:29.3879252Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3879323Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3879396Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3879474Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3879544Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3879626Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3879695Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3879764Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3879842Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3879914Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3879994Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3880065Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3880134Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3880213Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3880281Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3880361Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3880456Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3880639Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3880709Z return mod(*inputs) 2025-09-07T08:11:29.3880954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3881031Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3881276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3881389Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3881604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3881680Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3881930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3882028Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3882291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3882380Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3882646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3882775Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3882780Z 2025-09-07T08:11:29.3882875Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3883063Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3883126Z return mod(*inputs) 2025-09-07T08:11:29.3883369Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3883445Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3883691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3883765Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3883970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3884054Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3884295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3884378Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3884624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3884715Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3884986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3885088Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3885091Z 2025-09-07T08:11:29.3885163Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3885239Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3885333Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3885525Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3885586Z return mod(*inputs) 2025-09-07T08:11:29.3885826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3885903Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3886145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3886222Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3886424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3886505Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3886744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:11:29.3886818Z hidden_states = residual + hidden_states 2025-09-07T08:11:29.3886821Z 2025-09-07T08:11:29.3886925Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3886997Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3887075Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3887144Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3887213Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3887307Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3887374Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3887471Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3887540Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3887608Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3887687Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3887756Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3887834Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3887905Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3888000Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3888193Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3888255Z return mod(*inputs) 2025-09-07T08:11:29.3888506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3888574Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3888818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3888898Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3889099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3889178Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3889418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3889506Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3889756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3889854Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3890127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3890247Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3890251Z 2025-09-07T08:11:29.3890353Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3890534Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3890594Z return mod(*inputs) 2025-09-07T08:11:29.3890845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3890915Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3891167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3891233Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3891436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3891519Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3891763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3891852Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3892096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3892184Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3892489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3892589Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3892593Z 2025-09-07T08:11:29.3892674Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3892762Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3892840Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3892912Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3892997Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3893075Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3893144Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3893221Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3893293Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3893363Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3893441Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3893514Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3893583Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3893661Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3893731Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3893807Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3893905Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3894086Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3894155Z return mod(*inputs) 2025-09-07T08:11:29.3894398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3894474Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3894714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3894787Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3894994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3895066Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3895316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3895402Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3895650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3895738Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3896004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:11:29.3896127Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:11:29.3896131Z 2025-09-07T08:11:29.3896230Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3896416Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3896476Z return mod(*inputs) 2025-09-07T08:11:29.3896729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3896800Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3897040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3897115Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3897318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3897400Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3897644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:11:29.3897758Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:11:29.3898012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:11:29.3898100Z attn_output, attn_weights = attention_interface( 2025-09-07T08:11:29.3898394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:11:29.3898513Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:11:29.3898516Z 2025-09-07T08:11:29.3898594Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3898663Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3898757Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3898949Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3899009Z return mod(*inputs) 2025-09-07T08:11:29.3899260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:11:29.3899327Z encoder_outputs = self.encoder( 2025-09-07T08:11:29.3899569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:11:29.3899650Z layer_outputs = encoder_layer( 2025-09-07T08:11:29.3899852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:11:29.3899934Z return super().__call__(*args, **kwargs) 2025-09-07T08:11:29.3900175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:11:29.3900251Z hidden_states = residual + hidden_states 2025-09-07T08:11:29.3900255Z 2025-09-07T08:11:29.3900338Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3900413Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3900491Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3900560Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3900629Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3900705Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3900778Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3900881Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:29.3901062Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:29.3901124Z return mod(*inputs) 2025-09-07T08:11:29.3901374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1613, in forward 2025-09-07T08:11:29.3901457Z pooled_output = hidden_states.mean(dim=1) 2025-09-07T08:11:29.3901461Z 2025-09-07T08:11:29.3901539Z cudagraph partition due to non gpu ops 2025-09-07T08:11:29.3901612Z cudagraph partition due to non gpu ops 2025-09-07T08:11:52.7239738Z pass 2025-09-07T08:11:52.7258564Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:11:54.8725878Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:11:54.8727335Z import pynvml # type: ignore[import] 2025-09-07T08:11:56.5881778Z 2025-09-07T08:11:56.6785792Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:11:56.6786196Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:11:56.6786519Z cpu eval lennard_jones 2025-09-07T08:11:56.6797036Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:11:56.6816171Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:11:56.6825260Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:11:58.5607245Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:58.5607961Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:58.5608888Z return mod(*inputs) 2025-09-07T08:11:58.5609056Z 2025-09-07T08:11:58.5609230Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:58.5609937Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:58.5610489Z return mod(*inputs) 2025-09-07T08:11:58.5610652Z 2025-09-07T08:11:58.5610825Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:58.5611426Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:58.5611962Z return mod(*inputs) 2025-09-07T08:11:58.5612133Z 2025-09-07T08:11:58.5612311Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:58.5612911Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:58.5613445Z return mod(*inputs) 2025-09-07T08:11:58.5613601Z 2025-09-07T08:11:58.5613769Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:58.5614349Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:58.5614849Z return mod(*inputs) 2025-09-07T08:11:58.5615001Z 2025-09-07T08:11:58.5615159Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:58.5615712Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:58.5616191Z return mod(*inputs) 2025-09-07T08:11:58.5616340Z 2025-09-07T08:11:58.5616490Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:58.5617040Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:58.5617521Z return mod(*inputs) 2025-09-07T08:11:58.5617674Z 2025-09-07T08:11:58.5617852Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:58.5618458Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:58.5618946Z return mod(*inputs) 2025-09-07T08:11:58.5619086Z 2025-09-07T08:11:58.5619243Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:58.5619781Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:58.5620241Z return mod(*inputs) 2025-09-07T08:11:58.5620371Z 2025-09-07T08:11:58.5620494Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:58.5620984Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:58.5621433Z return mod(*inputs) 2025-09-07T08:11:58.5621572Z 2025-09-07T08:11:58.5621732Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:58.5622219Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:58.5622649Z return mod(*inputs) 2025-09-07T08:11:58.5622791Z 2025-09-07T08:11:58.5622911Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:58.5623336Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:11:58.5623696Z return mod(*inputs) 2025-09-07T08:11:58.5623817Z 2025-09-07T08:12:05.6543279Z pass 2025-09-07T08:12:05.6543714Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:06.7324401Z accuracy pass_rate=91.30% 2025-09-07T08:12:06.7324758Z calls_captured gmean=0.00x mean=420.913x 2025-09-07T08:12:06.7326803Z unique_graphs gmean=0.00x mean=3.826x 2025-09-07T08:12:06.7329109Z graph_breaks gmean=0.00x mean=2.174x 2025-09-07T08:12:06.7331844Z unique_graph_breaks gmean=0.00x mean=0.565x 2025-09-07T08:12:06.7333879Z autograd_captures gmean=0.00x mean=0.000x 2025-09-07T08:12:06.7336012Z autograd_compiles gmean=0.00x mean=0.000x 2025-09-07T08:12:06.7338243Z cudagraph_skips gmean=0.00x mean=2.957x 2025-09-07T08:12:06.7339010Z compilation_latency mean=23.066 seconds 2025-09-07T08:12:07.1216735Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true == *cppwrapper-true* ]] 2025-09-07T08:12:07.1217623Z + TORCHINDUCTOR_CPP_WRAPPER=1 2025-09-07T08:12:07.1218507Z + taskset -c 0-93 python benchmarks/dynamo/torchbench.py --accuracy --no-translation-validation --freezing --inference --amp --backend inductor --disable-cudagraphs --device cpu --total-partitions 4 --partition-id 1 --output /var/lib/jenkins/workspace/test/test-reports/inductor_cpp_wrapper_torchbench_amp_inference_cpu_x86_accuracy.csv 2025-09-07T08:12:07.4436858Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:12:07.4438204Z import pynvml # type: ignore[import] 2025-09-07T08:12:09.5120116Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:12:09.5121524Z import pynvml # type: ignore[import] 2025-09-07T08:12:11.2291268Z 2025-09-07T08:12:16.3229651Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:12:16.3231866Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:12:16.3233263Z cpu eval dlrm 2025-09-07T08:12:16.5204339Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:16.5700463Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:16.6181763Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:29.8141554Z pass 2025-09-07T08:12:29.8142067Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:31.3732757Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:12:31.3734171Z import pynvml # type: ignore[import] 2025-09-07T08:12:33.0930987Z 2025-09-07T08:12:35.9825918Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:12:35.9826240Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:12:35.9826617Z cpu eval doctr_det_predictor 2025-09-07T08:12:46.4479867Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:47.2950103Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:48.1392203Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:16.5702638Z pass 2025-09-07T08:13:16.5703078Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:18.6421448Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:13:18.6422866Z import pynvml # type: ignore[import] 2025-09-07T08:13:20.3658019Z 2025-09-07T08:13:22.4934985Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:13:22.4935296Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:13:22.4935539Z cpu eval doctr_reco_predictor 2025-09-07T08:13:22.6676154Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:22.6936087Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:22.7158430Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:36.2960288Z pass 2025-09-07T08:13:36.2961498Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:37.8815267Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:13:37.8816666Z import pynvml # type: ignore[import] 2025-09-07T08:13:39.5991266Z 2025-09-07T08:13:39.6435898Z loading model: 0it [00:00, ?it/s]Gym has been unmaintained since 2022 and does not support NumPy 2.0 amongst other critical functionality. 2025-09-07T08:13:39.6436554Z Please upgrade to Gymnasium, the maintained drop-in replacement of Gym, or contact the authors of your software and request that they upgrade. 2025-09-07T08:13:39.6437129Z Users of this version of Gym should be able to simply replace 'import gym' with 'import gymnasium as gym' in the vast majority of cases. 2025-09-07T08:13:39.6437685Z See the migration guide at https://gymnasium.farama.org/introduction/migration_guide/ for additional information. 2025-09-07T08:13:40.1293194Z 2025-09-07T08:13:40.1293754Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:13:40.1294179Z cpu eval drq 2025-09-07T08:13:40.1347305Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:40.1391972Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:40.1429646Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:52.3920140Z pass 2025-09-07T08:13:52.3920601Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:53.9543447Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:13:53.9544861Z import pynvml # type: ignore[import] 2025-09-07T08:13:55.6711331Z 2025-09-07T08:13:56.7705871Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:13:56.7706190Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:13:56.7706473Z cpu eval fastNLP_Bert 2025-09-07T08:13:57.2632777Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:57.4867695Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:57.7086675Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:57.7865658Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T08:13:57.7866257Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T08:13:57.7866716Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] or: 2025-09-07T08:13:57.7867522Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T08:13:57.7868076Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] to include these operations in the captured graph. 2025-09-07T08:13:57.7868599Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:13:57.7869102Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break: from user code at: 2025-09-07T08:13:57.7869775Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:13:57.7870368Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] return mod(*inputs) 2025-09-07T08:13:57.7870968Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/models/bert.py", line 265, in forward 2025-09-07T08:13:57.7871586Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] sequence_output = self.bert(words) 2025-09-07T08:13:57.7872233Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 137, in forward 2025-09-07T08:13:57.7872864Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] outputs = self.model(words) 2025-09-07T08:13:57.7873506Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 445, in forward 2025-09-07T08:13:57.7874648Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] max_word_piece_length = batch_word_pieces_length.sum(dim=-1).max().item() # 表示word piece的长度(包括padding) 2025-09-07T08:13:57.7875189Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:13:57.7875536Z W0907 08:13:57.785297 58204 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:14:21.3090239Z pass 2025-09-07T08:14:21.3090738Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:14:23.1744450Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:14:23.1745916Z import pynvml # type: ignore[import] 2025-09-07T08:14:24.8899177Z 2025-09-07T08:14:25.1920121Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:14:25.1920508Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:14:25.1920845Z cpu eval functorch_dp_cifar10 2025-09-07T08:14:25.2153824Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:14:25.2295708Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:14:25.2404488Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:14:40.6177471Z pass 2025-09-07T08:14:40.6177924Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:14:42.1836862Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:14:42.1838308Z import pynvml # type: ignore[import] 2025-09-07T08:14:43.9003360Z 2025-09-07T08:14:44.0225265Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:14:44.0226065Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:14:44.0226404Z cpu eval functorch_maml_omniglot 2025-09-07T08:14:44.0260547Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:14:44.0299525Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:14:44.0325151Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:14:54.5818636Z pass 2025-09-07T08:14:54.5819057Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:14:56.0338866Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:14:56.0340302Z import pynvml # type: ignore[import] 2025-09-07T08:14:57.7502692Z 2025-09-07T08:14:59.0403423Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:14:59.0403791Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:14:59.0404073Z cpu eval hf_Albert 2025-09-07T08:14:59.5107375Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:14:59.5948373Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:14:59.6779506Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:15:17.7597520Z pass 2025-09-07T08:15:17.7597937Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:15:19.5816897Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:15:19.5818313Z import pynvml # type: ignore[import] 2025-09-07T08:15:21.3060781Z 2025-09-07T08:15:24.6935119Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:15:24.6935554Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:15:24.6935869Z cpu eval hf_Bart 2025-09-07T08:15:25.3771128Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:15:25.5603460Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:15:25.7170568Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:15:50.7926016Z pass 2025-09-07T08:15:50.7926443Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:15:52.8612686Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:15:52.8614119Z import pynvml # type: ignore[import] 2025-09-07T08:15:54.5762375Z 2025-09-07T08:15:56.9456251Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:15:56.9456595Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:15:56.9456837Z cpu eval hf_Bert 2025-09-07T08:15:57.4637658Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:15:57.5752935Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:15:57.6841751Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:17.0931980Z pass 2025-09-07T08:16:17.0932765Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:19.0432816Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:16:19.0434659Z import pynvml # type: ignore[import] 2025-09-07T08:16:20.7783713Z 2025-09-07T08:16:24.6409443Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:16:24.6409768Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:16:24.6409996Z cpu eval hf_Bert_large 2025-09-07T08:16:25.8866870Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:26.1463416Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:26.4050144Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:56.7204748Z pass 2025-09-07T08:16:56.7205248Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:58.8892627Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:16:58.8894103Z import pynvml # type: ignore[import] 2025-09-07T08:17:00.6018136Z 2025-09-07T08:17:01.7472980Z loading model: 0it [00:00, ?it/s]BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:17:01.7474198Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:17:01.7474936Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:17:01.7475549Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:17:01.7476691Z WARNING:transformers.modeling_utils:BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:17:01.7477773Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:17:01.7478482Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:17:01.7479064Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:17:03.0961432Z Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:17:03.0962716Z WARNING:transformers.models.big_bird.modeling_big_bird:Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:17:04.1442705Z 2025-09-07T08:17:04.1443145Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:17:04.1444332Z cpu eval hf_BigBird 2025-09-07T08:17:05.5290406Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:17:07.0090921Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:17:08.4899608Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:06.1701525Z W0907 08:18:06.169423 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] torch._dynamo hit config.recompile_limit (8) 2025-09-07T08:18:06.1702729Z W0907 08:18:06.169423 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] function: 'forward' (/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py:417) 2025-09-07T08:18:06.1704490Z W0907 08:18:06.169423 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] last reason: 10/7: self.seed == 7 # context_layer, attention_probs = self.bigbird_block_sparse_attention( # transformers/models/big_bird/modeling_big_bird.py:455 in forward (HINT: torch.compile considers integer attributes of the nn.Module to be static. If you are observing recompilation, you might want to make this integer dynamic using torch._dynamo.config.allow_unspec_int_on_nn_module = True, or convert this integer into a tensor.) 2025-09-07T08:18:06.1706018Z W0907 08:18:06.169423 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] To log all recompilation reasons, use TORCH_LOGS="recompiles". 2025-09-07T08:18:06.1706805Z W0907 08:18:06.169423 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html 2025-09-07T08:18:06.2535133Z W0907 08:18:06.252176 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] torch._dynamo hit config.recompile_limit (8) 2025-09-07T08:18:06.2536086Z W0907 08:18:06.252176 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] function: 'torch_dynamo_resume_in_forward_at_1381' (/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py:1381) 2025-09-07T08:18:06.2537113Z W0907 08:18:06.252176 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] last reason: 17/7: ___check_obj_id(self._modules['output']._modules['dense']._parameters['bias'], 139800347385104) # return F.linear(input, self.weight, self.bias) # nn/modules/linear.py:134 in forward 2025-09-07T08:18:06.2537927Z W0907 08:18:06.252176 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] To log all recompilation reasons, use TORCH_LOGS="recompiles". 2025-09-07T08:18:06.2538603Z W0907 08:18:06.252176 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html 2025-09-07T08:18:06.2570348Z W0907 08:18:06.256587 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] torch._dynamo hit config.recompile_limit (8) 2025-09-07T08:18:06.2571342Z W0907 08:18:06.256587 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] function: 'torch_dynamo_resume_in_forward_at_1468' (/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py:1468) 2025-09-07T08:18:06.2572674Z W0907 08:18:06.256587 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] last reason: 18/7: ___check_obj_id(self._modules['output']._modules['dense']._parameters['bias'], 139800347385024) # return F.linear(input, self.weight, self.bias) # nn/modules/linear.py:134 in forward 2025-09-07T08:18:06.2573738Z W0907 08:18:06.256587 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] To log all recompilation reasons, use TORCH_LOGS="recompiles". 2025-09-07T08:18:06.2575128Z W0907 08:18:06.256587 66271 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html 2025-09-07T08:18:10.5016656Z pass 2025-09-07T08:18:10.5017143Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:12.7918430Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:18:12.7920461Z import pynvml # type: ignore[import] 2025-09-07T08:18:14.5148402Z 2025-09-07T08:18:16.6392091Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:18:16.6392392Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:18:16.6392635Z cpu eval hf_DistilBert 2025-09-07T08:18:16.9580562Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:17.0221729Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:17.0845486Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:32.1321677Z pass 2025-09-07T08:18:32.1322107Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:34.0041722Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:18:34.0043151Z import pynvml # type: ignore[import] 2025-09-07T08:18:35.7222421Z 2025-09-07T08:18:39.4896517Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:18:39.4899071Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:18:39.4900710Z cpu eval hf_GPT2 2025-09-07T08:18:39.7363965Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:39.8117527Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:39.8864185Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:59.8112870Z pass 2025-09-07T08:18:59.8113341Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:01.8740409Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:19:01.8741821Z import pynvml # type: ignore[import] 2025-09-07T08:19:03.5945349Z 2025-09-07T08:19:14.3140067Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:19:14.3140536Z loading model: 0it [00:10, ?it/s] 2025-09-07T08:19:14.3140862Z cpu eval hf_GPT2_large 2025-09-07T08:19:14.6880905Z pass_due_to_skip 2025-09-07T08:19:14.6881456Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:16.2236713Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:19:16.2238176Z import pynvml # type: ignore[import] 2025-09-07T08:19:17.9391172Z 2025-09-07T08:19:21.0071282Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:19:21.0071613Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:19:21.0071892Z cpu eval hf_Longformer 2025-09-07T08:19:22.0139530Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:23.1661025Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:24.3160879Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:24.4690535Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T08:19:24.4691574Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T08:19:24.4692170Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] or: 2025-09-07T08:19:24.4692617Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T08:19:24.4693151Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] to include these operations in the captured graph. 2025-09-07T08:19:24.4693582Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:19:24.4694002Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break: from user code at: 2025-09-07T08:19:24.4694634Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:19:24.4695219Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] return mod(**inputs) 2025-09-07T08:19:24.4695899Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1703, in forward 2025-09-07T08:19:24.4696580Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] outputs = self.longformer( 2025-09-07T08:19:24.4697256Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1600, in forward 2025-09-07T08:19:24.4697925Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] encoder_outputs = self.encoder( 2025-09-07T08:19:24.4698607Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1244, in forward 2025-09-07T08:19:24.4699325Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] is_global_attn = is_index_global_attn.flatten().any().item() 2025-09-07T08:19:24.4699780Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:19:24.4700135Z W0907 08:19:24.467659 73246 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:20:14.8096039Z pass 2025-09-07T08:20:14.8096610Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:20:17.2527403Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:20:17.2528859Z import pynvml # type: ignore[import] 2025-09-07T08:20:18.9765782Z 2025-09-07T08:20:20.0390122Z loading model: 0it [00:00, ?it/s]Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T08:20:20.0394431Z WARNING:transformers.models.reformer.modeling_reformer:Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T08:20:20.4410288Z 2025-09-07T08:20:20.4410920Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:20:20.4411937Z cpu eval hf_Reformer 2025-09-07T08:20:20.7852291Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:20:21.5567171Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:20:22.3263145Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:20:32.0495995Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T08:20:32.0496720Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T08:20:32.0497264Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] or: 2025-09-07T08:20:32.0497804Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T08:20:32.0498418Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] to include these operations in the captured graph. 2025-09-07T08:20:32.0498938Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T08:20:32.0499430Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] Graph break: from user code at: 2025-09-07T08:20:32.0500325Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2132, in torch_dynamo_resume_in_forward_at_2104 2025-09-07T08:20:32.0501222Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] embedding_output = self.embeddings( 2025-09-07T08:20:32.0502023Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 369, in forward 2025-09-07T08:20:32.0502863Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] position_embeddings = self.position_embeddings(position_ids) 2025-09-07T08:20:32.0503698Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 293, in forward 2025-09-07T08:20:32.0504484Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] max_position_id = position_ids.max().item() 2025-09-07T08:20:32.0504994Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T08:20:32.0505398Z W0907 08:20:32.048896 74689 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T08:20:49.4253146Z pass 2025-09-07T08:20:49.4253641Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:20:51.2807931Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:20:51.2809352Z import pynvml # type: ignore[import] 2025-09-07T08:20:53.0019294Z 2025-09-07T08:20:57.7882707Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:20:57.7883023Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:20:57.7883725Z cpu eval hf_Roberta_base 2025-09-07T08:20:59.1602221Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:20:59.4059542Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:20:59.6304263Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:21:22.0134399Z pass 2025-09-07T08:21:22.0134834Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:21:24.3112455Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:21:24.3113902Z import pynvml # type: ignore[import] 2025-09-07T08:21:26.0323954Z 2025-09-07T08:21:28.2658638Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:21:28.2659688Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:21:28.2660160Z cpu eval hf_T5 2025-09-07T08:21:28.7326029Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:21:28.9658779Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:21:29.1964681Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:21:50.2527469Z pass 2025-09-07T08:21:50.2527905Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:21:52.2312677Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:21:52.2314112Z import pynvml # type: ignore[import] 2025-09-07T08:21:53.9466034Z 2025-09-07T08:22:03.9479658Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:22:03.9481159Z loading model: 0it [00:10, ?it/s] 2025-09-07T08:22:03.9482191Z cpu eval hf_T5_base 2025-09-07T08:22:16.9180394Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:22:24.7772359Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:22:32.6165036Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:23:15.3838048Z pass 2025-09-07T08:23:15.3838467Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:23:18.7137596Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:23:18.7139020Z import pynvml # type: ignore[import] 2025-09-07T08:23:20.4339860Z 2025-09-07T08:23:29.9646223Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:23:29.9646674Z loading model: 0it [00:09, ?it/s] 2025-09-07T08:23:29.9648804Z cpu eval hf_T5_large 2025-09-07T08:23:30.3580639Z pass_due_to_skip 2025-09-07T08:23:30.3581187Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:23:31.8141465Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:23:31.8143459Z import pynvml # type: ignore[import] 2025-09-07T08:23:33.5306211Z 2025-09-07T08:23:37.8537224Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:23:37.8537537Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:23:37.8537782Z cpu eval hf_distil_whisper 2025-09-07T08:23:42.3495654Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:23:43.0947072Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:23:43.8388863Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:24:17.0496420Z pass 2025-09-07T08:24:17.0496820Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:24:19.3200379Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:24:19.3201777Z import pynvml # type: ignore[import] 2025-09-07T08:24:21.0339967Z 2025-09-07T08:24:21.1244312Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:24:21.1244698Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:24:21.1245045Z cpu eval lennard_jones 2025-09-07T08:24:21.1254987Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:24:21.1273327Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:24:21.1282035Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:24:30.5545912Z pass 2025-09-07T08:24:30.5546355Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:24:31.6304984Z accuracy pass_rate=91.30% 2025-09-07T08:24:31.6305340Z calls_captured gmean=0.00x mean=415.739x 2025-09-07T08:24:31.6305592Z unique_graphs gmean=0.00x mean=3.783x 2025-09-07T08:24:31.6306503Z graph_breaks gmean=0.00x mean=2.174x 2025-09-07T08:24:31.6308779Z unique_graph_breaks gmean=0.00x mean=0.565x 2025-09-07T08:24:31.6311279Z autograd_captures gmean=0.00x mean=0.000x 2025-09-07T08:24:31.6313436Z autograd_compiles gmean=0.00x mean=0.000x 2025-09-07T08:24:31.6315712Z cudagraph_skips gmean=0.00x mean=0.000x 2025-09-07T08:24:31.6316481Z compilation_latency mean=21.693 seconds 2025-09-07T08:24:32.0153112Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true == *freezing_cudagraphs-true* ]] 2025-09-07T08:24:32.0153981Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true == *freeze_autotune_cudagraphs-true* ]] 2025-09-07T08:24:32.0154793Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true == *aotinductor-true* ]] 2025-09-07T08:24:32.0155284Z + [[ inference == \i\n\f\e\r\e\n\c\e ]] 2025-09-07T08:24:32.0155494Z + [[ accuracy == \a\c\c\u\r\a\c\y ]] 2025-09-07T08:24:32.0156314Z + taskset -c 0-93 python benchmarks/dynamo/torchbench.py --accuracy --no-translation-validation --freezing --inference --amp --export --disable-cudagraphs --device cpu --total-partitions 4 --partition-id 1 --output /var/lib/jenkins/workspace/test/test-reports/inductor_export_torchbench_amp_inference_cpu_x86_accuracy.csv 2025-09-07T08:24:32.3553708Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:24:32.3555236Z import pynvml # type: ignore[import] 2025-09-07T08:24:34.4144985Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:24:34.4146423Z import pynvml # type: ignore[import] 2025-09-07T08:24:36.1436514Z 2025-09-07T08:24:41.2325567Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:24:41.2327802Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:24:41.2330401Z cpu eval dlrm 2025-09-07T08:24:41.4264008Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:24:41.4759017Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:24:41.5238127Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:24:49.6686668Z pass 2025-09-07T08:24:49.6687125Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:24:50.6628193Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:24:50.6629638Z import pynvml # type: ignore[import] 2025-09-07T08:24:52.3813144Z 2025-09-07T08:24:55.2601571Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:24:55.2601866Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:24:55.2602117Z cpu eval doctr_det_predictor 2025-09-07T08:25:05.6767172Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:06.5246529Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:07.3772841Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:08.4523736Z ERROR:common: 2025-09-07T08:25:08.4523994Z Traceback (most recent call last): 2025-09-07T08:25:08.4524364Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 2320, in check_accuracy 2025-09-07T08:25:08.4524729Z optimized_model_iter_fn = optimize_ctx( 2025-09-07T08:25:08.4525066Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 1523, in export 2025-09-07T08:25:08.4525380Z ep = torch.export.export( 2025-09-07T08:25:08.4525719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 311, in export 2025-09-07T08:25:08.4526030Z raise e 2025-09-07T08:25:08.4526292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 277, in export 2025-09-07T08:25:08.4526606Z return _export( 2025-09-07T08:25:08.4526891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:25:08.4527206Z raise e 2025-09-07T08:25:08.4527472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:25:08.4527778Z ep = fn(*args, **kwargs) 2025-09-07T08:25:08.4528104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:25:08.4528446Z return fn(*args, **kwargs) 2025-09-07T08:25:08.4528745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2255, in _export 2025-09-07T08:25:08.4529053Z ep = _export_for_training( 2025-09-07T08:25:08.4529347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:25:08.4529643Z raise e 2025-09-07T08:25:08.4529904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:25:08.4530209Z ep = fn(*args, **kwargs) 2025-09-07T08:25:08.4530886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:25:08.4531229Z return fn(*args, **kwargs) 2025-09-07T08:25:08.4531561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2071, in _export_for_training 2025-09-07T08:25:08.4531975Z export_artifact = export_func( 2025-09-07T08:25:08.4532298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1415, in _strict_export 2025-09-07T08:25:08.4532680Z gm_torch_level = _export_to_torch_ir( 2025-09-07T08:25:08.4533026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 812, in _export_to_torch_ir 2025-09-07T08:25:08.4533374Z gm_torch_level, _ = torch._dynamo.export( 2025-09-07T08:25:08.4533708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 2002, in inner 2025-09-07T08:25:08.4534064Z result_traced = opt_f(*args, **kwargs) 2025-09-07T08:25:08.4534419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 414, in __call__ 2025-09-07T08:25:08.4534755Z return super().__call__(*args, **kwargs) 2025-09-07T08:25:08.4535125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl 2025-09-07T08:25:08.4535512Z return self._call_impl(*args, **kwargs) 2025-09-07T08:25:08.4535877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:25:08.4536229Z return forward_call(*args, **kwargs) 2025-09-07T08:25:08.4536577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 841, in compile_wrapper 2025-09-07T08:25:08.4536993Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-09-07T08:25:08.4537314Z torch._dynamo.exc.Unsupported: Attempted to call function marked as skipped 2025-09-07T08:25:08.4537949Z Explanation: Dynamo does not know how to trace the builtin `.morphologyEx.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind). 2025-09-07T08:25:08.4538688Z Hint: If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround. 2025-09-07T08:25:08.4539510Z Hint: If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`. 2025-09-07T08:25:08.4540068Z 2025-09-07T08:25:08.4540266Z Developer debug context: module: , qualname: morphologyEx, skip reason: 2025-09-07T08:25:08.4540505Z 2025-09-07T08:25:08.4540765Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0007.html 2025-09-07T08:25:08.4541054Z 2025-09-07T08:25:08.4541123Z from user code: 2025-09-07T08:25:08.4541506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 211, in forward 2025-09-07T08:25:08.4542013Z for preds in self.postprocessor(prob_map.detach().cpu().permute((0, 2, 3, 1)).numpy()) 2025-09-07T08:25:08.4542437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/core.py", line 90, in __call__ 2025-09-07T08:25:08.4542762Z bin_map = [ 2025-09-07T08:25:08.4543058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/core.py", line 91, in 2025-09-07T08:25:08.4543383Z [ 2025-09-07T08:25:08.4543668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/core.py", line 92, in 2025-09-07T08:25:08.4544070Z cv2.morphologyEx(bmap[..., idx], cv2.MORPH_OPEN, self._opening_kernel) 2025-09-07T08:25:08.4544295Z 2025-09-07T08:25:08.4544624Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-09-07T08:25:08.4544991Z 2025-09-07T08:25:08.4545133Z TorchDynamo optimized model failed to run because of following error 2025-09-07T08:25:08.6857383Z fail_to_run 2025-09-07T08:25:08.6857839Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:09.8901855Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:25:09.8903265Z import pynvml # type: ignore[import] 2025-09-07T08:25:11.6066663Z 2025-09-07T08:25:13.7429608Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:25:13.7429952Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:25:13.7430192Z cpu eval doctr_reco_predictor 2025-09-07T08:25:13.9109733Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:13.9366918Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:13.9587499Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:14.5108706Z ERROR:common: 2025-09-07T08:25:14.5109001Z Traceback (most recent call last): 2025-09-07T08:25:14.5109469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/variables/tensor.py", line 1435, in evaluate_expr 2025-09-07T08:25:14.5109917Z return guard_scalar(self.sym_num) 2025-09-07T08:25:14.5110334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/symbolic_shapes.py", line 1535, in guard_scalar 2025-09-07T08:25:14.5110741Z return guard_int(a) 2025-09-07T08:25:14.5111102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/symbolic_shapes.py", line 1732, in guard_int 2025-09-07T08:25:14.5111519Z return a.node.guard_int("", 0) # NB: uses Python backtrace 2025-09-07T08:25:14.5111926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/sym_node.py", line 518, in guard_int 2025-09-07T08:25:14.5112278Z r = self.evaluate() 2025-09-07T08:25:14.5112610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/sym_node.py", line 512, in evaluate 2025-09-07T08:25:14.5113011Z return self.shape_env.evaluate_sym_node(self, size_oblivious) 2025-09-07T08:25:14.5113442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/symbolic_shapes.py", line 7239, in evaluate_sym_node 2025-09-07T08:25:14.5113836Z return self.evaluate_expr( 2025-09-07T08:25:14.5114213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/symbolic_shapes.py", line 7339, in evaluate_expr 2025-09-07T08:25:14.5114597Z return self._inner_evaluate_expr( 2025-09-07T08:25:14.5114949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/recording.py", line 272, in wrapper 2025-09-07T08:25:14.5115298Z return retlog(fn(*args, **kwargs)) 2025-09-07T08:25:14.5115699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/symbolic_shapes.py", line 7362, in _inner_evaluate_expr 2025-09-07T08:25:14.5116089Z return self._evaluate_expr( 2025-09-07T08:25:14.5116455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/symbolic_shapes.py", line 7580, in _evaluate_expr 2025-09-07T08:25:14.5116842Z raise self._make_data_dependent_error( 2025-09-07T08:25:14.5117367Z torch.fx.experimental.symbolic_shapes.GuardOnDataDependentSymNode: Could not extract specialized integer from data-dependent expression u0 (unhinted: u0). (Size-like symbols: none) 2025-09-07T08:25:14.5118103Z 2025-09-07T08:25:14.5118527Z Caused by: decode_sequence([k for k, _ in groupby(seq.tolist()) if k != blank], vocab) # doctr/models/recognition/crnn/pytorch.py:79 in (_dynamo/variables/tensor.py:1435 in evaluate_expr) 2025-09-07T08:25:14.5119058Z For more information, run with TORCH_LOGS="dynamic" 2025-09-07T08:25:14.5119475Z For extended logs when we create symbols, also add TORCHDYNAMO_EXTENDED_DEBUG_CREATE_SYMBOL="u0" 2025-09-07T08:25:14.5119938Z If you suspect the guard was triggered from C++, add TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:25:14.5120487Z For more debugging help, see https://docs.google.com/document/d/1HSuTTVvYH1pTew89Rtpeu84Ht3nQEFTYhAX3Ypa_xJs/edit?usp=sharing 2025-09-07T08:25:14.5120828Z 2025-09-07T08:25:14.5120902Z User Stack (most recent call last): 2025-09-07T08:25:14.5121106Z (snipped, see stack below for prefix) 2025-09-07T08:25:14.5121484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 224, in forward 2025-09-07T08:25:14.5121864Z out["preds"] = self.postprocessor(logits) 2025-09-07T08:25:14.5122230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 99, in __call__ 2025-09-07T08:25:14.5122673Z return self.ctc_best_path(logits=logits, vocab=self.vocab, blank=len(self.vocab)) 2025-09-07T08:25:14.5123128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 78, in ctc_best_path 2025-09-07T08:25:14.5123491Z words = [ 2025-09-07T08:25:14.5123815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 79, in 2025-09-07T08:25:14.5124233Z decode_sequence([k for k, _ in groupby(seq.tolist()) if k != blank], vocab) 2025-09-07T08:25:14.5124424Z 2025-09-07T08:25:14.5124535Z For C++ stack trace, run with TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:25:14.5124706Z 2025-09-07T08:25:14.5124835Z During handling of the above exception, another exception occurred: 2025-09-07T08:25:14.5125008Z 2025-09-07T08:25:14.5125392Z torch._dynamo.exc.UserError: Consider annotating your code using torch._check*(). Could not extract specialized integer from data-dependent expression u0 (unhinted: u0). (Size-like symbols: none) 2025-09-07T08:25:14.5125817Z 2025-09-07T08:25:14.5126162Z Caused by: decode_sequence([k for k, _ in groupby(seq.tolist()) if k != blank], vocab) # doctr/models/recognition/crnn/pytorch.py:79 in (_dynamo/variables/tensor.py:1435 in evaluate_expr) 2025-09-07T08:25:14.5126643Z For more information, run with TORCH_LOGS="dynamic" 2025-09-07T08:25:14.5126953Z For extended logs when we create symbols, also add TORCHDYNAMO_EXTENDED_DEBUG_CREATE_SYMBOL="u0" 2025-09-07T08:25:14.5127328Z If you suspect the guard was triggered from C++, add TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:25:14.5127792Z For more debugging help, see https://docs.google.com/document/d/1HSuTTVvYH1pTew89Rtpeu84Ht3nQEFTYhAX3Ypa_xJs/edit?usp=sharing 2025-09-07T08:25:14.5128101Z 2025-09-07T08:25:14.5128184Z User Stack (most recent call last): 2025-09-07T08:25:14.5128385Z (snipped, see stack below for prefix) 2025-09-07T08:25:14.5128743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 224, in forward 2025-09-07T08:25:14.5129120Z out["preds"] = self.postprocessor(logits) 2025-09-07T08:25:14.5129496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 99, in __call__ 2025-09-07T08:25:14.5129934Z return self.ctc_best_path(logits=logits, vocab=self.vocab, blank=len(self.vocab)) 2025-09-07T08:25:14.5130371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 78, in ctc_best_path 2025-09-07T08:25:14.5130734Z words = [ 2025-09-07T08:25:14.5131096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 79, in 2025-09-07T08:25:14.5131522Z decode_sequence([k for k, _ in groupby(seq.tolist()) if k != blank], vocab) 2025-09-07T08:25:14.5131704Z 2025-09-07T08:25:14.5131822Z For C++ stack trace, run with TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:25:14.5132239Z For more information about this error, see: https://pytorch.org/docs/main/generated/exportdb/index.html#constrain-as-size-example 2025-09-07T08:25:14.5132570Z 2025-09-07T08:25:14.5132721Z The above exception was the direct cause of the following exception: 2025-09-07T08:25:14.5132907Z 2025-09-07T08:25:14.5132983Z Traceback (most recent call last): 2025-09-07T08:25:14.5133295Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 2320, in check_accuracy 2025-09-07T08:25:14.5133612Z optimized_model_iter_fn = optimize_ctx( 2025-09-07T08:25:14.5133901Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 1523, in export 2025-09-07T08:25:14.5134194Z ep = torch.export.export( 2025-09-07T08:25:14.5134509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 311, in export 2025-09-07T08:25:14.5134812Z raise e 2025-09-07T08:25:14.5135072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 277, in export 2025-09-07T08:25:14.5135375Z return _export( 2025-09-07T08:25:14.5135652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:25:14.5135949Z raise e 2025-09-07T08:25:14.5136208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:25:14.5136505Z ep = fn(*args, **kwargs) 2025-09-07T08:25:14.5136826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:25:14.5137180Z return fn(*args, **kwargs) 2025-09-07T08:25:14.5137473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2255, in _export 2025-09-07T08:25:14.5137778Z ep = _export_for_training( 2025-09-07T08:25:14.5138075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:25:14.5138370Z raise e 2025-09-07T08:25:14.5138621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:25:14.5138924Z ep = fn(*args, **kwargs) 2025-09-07T08:25:14.5139244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:25:14.5139575Z return fn(*args, **kwargs) 2025-09-07T08:25:14.5139907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2071, in _export_for_training 2025-09-07T08:25:14.5140247Z export_artifact = export_func( 2025-09-07T08:25:14.5140576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1415, in _strict_export 2025-09-07T08:25:14.5140915Z gm_torch_level = _export_to_torch_ir( 2025-09-07T08:25:14.5141256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 812, in _export_to_torch_ir 2025-09-07T08:25:14.5141604Z gm_torch_level, _ = torch._dynamo.export( 2025-09-07T08:25:14.5141930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 2002, in inner 2025-09-07T08:25:14.5142252Z result_traced = opt_f(*args, **kwargs) 2025-09-07T08:25:14.5142583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 414, in __call__ 2025-09-07T08:25:14.5142908Z return super().__call__(*args, **kwargs) 2025-09-07T08:25:14.5143261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl 2025-09-07T08:25:14.5143620Z return self._call_impl(*args, **kwargs) 2025-09-07T08:25:14.5143991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:25:14.5144322Z return forward_call(*args, **kwargs) 2025-09-07T08:25:14.5144663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 841, in compile_wrapper 2025-09-07T08:25:14.5145045Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-09-07T08:25:14.5145407Z torch._dynamo.exc.Unsupported: Unexpected failure during itertools.groupby() iteration 2025-09-07T08:25:14.5145783Z Explanation: Unexpected failure in invoking function during groupby 2025-09-07T08:25:14.5146289Z Hint: It may be possible to write Dynamo tracing rules for this code. Please report an issue to PyTorch if you encounter this graph break often and it is causing performance issues. 2025-09-07T08:25:14.5146652Z 2025-09-07T08:25:14.5146882Z Developer debug context: call_function ItertoolsVariable() [ListVariable(length=32)] {} 2025-09-07T08:25:14.5147156Z 2025-09-07T08:25:14.5147403Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0115.html 2025-09-07T08:25:14.5147705Z 2025-09-07T08:25:14.5147783Z from user code: 2025-09-07T08:25:14.5148117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 224, in forward 2025-09-07T08:25:14.5148502Z out["preds"] = self.postprocessor(logits) 2025-09-07T08:25:14.5148876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 99, in __call__ 2025-09-07T08:25:14.5149315Z return self.ctc_best_path(logits=logits, vocab=self.vocab, blank=len(self.vocab)) 2025-09-07T08:25:14.5149759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 78, in ctc_best_path 2025-09-07T08:25:14.5150114Z words = [ 2025-09-07T08:25:14.5150437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 79, in 2025-09-07T08:25:14.5150866Z decode_sequence([k for k, _ in groupby(seq.tolist()) if k != blank], vocab) 2025-09-07T08:25:14.5151048Z 2025-09-07T08:25:14.5151381Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-09-07T08:25:14.5151745Z 2025-09-07T08:25:14.5151877Z TorchDynamo optimized model failed to run because of following error 2025-09-07T08:25:14.5629925Z fail_to_run 2025-09-07T08:25:15.5626758Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:15.5628402Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:25:15.5629916Z import pynvml # type: ignore[import] 2025-09-07T08:25:17.2822936Z 2025-09-07T08:25:17.3263898Z loading model: 0it [00:00, ?it/s]Gym has been unmaintained since 2022 and does not support NumPy 2.0 amongst other critical functionality. 2025-09-07T08:25:17.3264513Z Please upgrade to Gymnasium, the maintained drop-in replacement of Gym, or contact the authors of your software and request that they upgrade. 2025-09-07T08:25:17.3265105Z Users of this version of Gym should be able to simply replace 'import gym' with 'import gymnasium as gym' in the vast majority of cases. 2025-09-07T08:25:17.3265652Z See the migration guide at https://gymnasium.farama.org/introduction/migration_guide/ for additional information. 2025-09-07T08:25:17.8131076Z 2025-09-07T08:25:17.8131651Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:25:17.8132063Z cpu eval drq 2025-09-07T08:25:17.8186317Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:17.8233691Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:17.8269151Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:25.1188789Z pass 2025-09-07T08:25:25.1189511Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:26.0116421Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:25:26.0118143Z import pynvml # type: ignore[import] 2025-09-07T08:25:27.7315769Z 2025-09-07T08:25:28.8342479Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:25:28.8342810Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:25:28.8343059Z cpu eval fastNLP_Bert 2025-09-07T08:25:29.3329335Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:29.5566226Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:29.7795424Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:30.0016693Z ERROR:common: 2025-09-07T08:25:30.0016947Z Traceback (most recent call last): 2025-09-07T08:25:30.0017345Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 2320, in check_accuracy 2025-09-07T08:25:30.0017674Z optimized_model_iter_fn = optimize_ctx( 2025-09-07T08:25:30.0017988Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 1523, in export 2025-09-07T08:25:30.0018290Z ep = torch.export.export( 2025-09-07T08:25:30.0018621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 311, in export 2025-09-07T08:25:30.0018935Z raise e 2025-09-07T08:25:30.0019202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 277, in export 2025-09-07T08:25:30.0019507Z return _export( 2025-09-07T08:25:30.0019788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:25:30.0020087Z raise e 2025-09-07T08:25:30.0020350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:25:30.0020661Z ep = fn(*args, **kwargs) 2025-09-07T08:25:30.0020988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:25:30.0021317Z return fn(*args, **kwargs) 2025-09-07T08:25:30.0021607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2255, in _export 2025-09-07T08:25:30.0021917Z ep = _export_for_training( 2025-09-07T08:25:30.0022217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:25:30.0022512Z raise e 2025-09-07T08:25:30.0022767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:25:30.0023069Z ep = fn(*args, **kwargs) 2025-09-07T08:25:30.0023390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:25:30.0023721Z return fn(*args, **kwargs) 2025-09-07T08:25:30.0024049Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2071, in _export_for_training 2025-09-07T08:25:30.0024387Z export_artifact = export_func( 2025-09-07T08:25:30.0024715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1415, in _strict_export 2025-09-07T08:25:30.0025052Z gm_torch_level = _export_to_torch_ir( 2025-09-07T08:25:30.0025737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 812, in _export_to_torch_ir 2025-09-07T08:25:30.0026089Z gm_torch_level, _ = torch._dynamo.export( 2025-09-07T08:25:30.0026426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 2002, in inner 2025-09-07T08:25:30.0026801Z result_traced = opt_f(*args, **kwargs) 2025-09-07T08:25:30.0027129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 414, in __call__ 2025-09-07T08:25:30.0027503Z return super().__call__(*args, **kwargs) 2025-09-07T08:25:30.0027859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl 2025-09-07T08:25:30.0028218Z return self._call_impl(*args, **kwargs) 2025-09-07T08:25:30.0028561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:25:30.0028899Z return forward_call(*args, **kwargs) 2025-09-07T08:25:30.0029236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 841, in compile_wrapper 2025-09-07T08:25:30.0029631Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-09-07T08:25:30.0030174Z torch._dynamo.exc.UserError: Consider annotating your code using torch._check*(). Could not guard on data-dependent expression u0 + 2 > 512 (unhinted: u0 + 2 > 512). (Size-like symbols: none) 2025-09-07T08:25:30.0030591Z 2025-09-07T08:25:30.0031125Z consider using data-dependent friendly APIs such as guard_or_false, guard_or_true and statically_known_trueCaused by: if max_word_piece_length + 2 > self._max_position_embeddings: # fastNLP/embeddings/bert_embedding.py:446 in forward (_dynamo/variables/tensor.py:1435 in evaluate_expr) 2025-09-07T08:25:30.0031825Z For more information, run with TORCH_LOGS="dynamic" 2025-09-07T08:25:30.0032149Z For extended logs when we create symbols, also add TORCHDYNAMO_EXTENDED_DEBUG_CREATE_SYMBOL="u0" 2025-09-07T08:25:30.0032526Z If you suspect the guard was triggered from C++, add TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:25:30.0033031Z For more debugging help, see https://docs.google.com/document/d/1HSuTTVvYH1pTew89Rtpeu84Ht3nQEFTYhAX3Ypa_xJs/edit?usp=sharing 2025-09-07T08:25:30.0033359Z 2025-09-07T08:25:30.0033436Z User Stack (most recent call last): 2025-09-07T08:25:30.0033639Z (snipped, see stack below for prefix) 2025-09-07T08:25:30.0033965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/models/bert.py", line 265, in forward 2025-09-07T08:25:30.0034280Z sequence_output = self.bert(words) 2025-09-07T08:25:30.0034617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:25:30.0034950Z return forward_call(*args, **kwargs) 2025-09-07T08:25:30.0035316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 137, in forward 2025-09-07T08:25:30.0035663Z outputs = self.model(words) 2025-09-07T08:25:30.0035980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:25:30.0036317Z return forward_call(*args, **kwargs) 2025-09-07T08:25:30.0036660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 446, in forward 2025-09-07T08:25:30.0037050Z if max_word_piece_length + 2 > self._max_position_embeddings: 2025-09-07T08:25:30.0037212Z 2025-09-07T08:25:30.0037333Z For C++ stack trace, run with TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:25:30.0037759Z For more information about this error, see: https://pytorch.org/docs/main/generated/exportdb/index.html#constrain-as-size-example 2025-09-07T08:25:30.0038078Z 2025-09-07T08:25:30.0038137Z from user code: 2025-09-07T08:25:30.0038418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/models/bert.py", line 265, in forward 2025-09-07T08:25:30.0038730Z sequence_output = self.bert(words) 2025-09-07T08:25:30.0039105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:25:30.0039426Z return forward_call(*args, **kwargs) 2025-09-07T08:25:30.0039765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 137, in forward 2025-09-07T08:25:30.0040126Z outputs = self.model(words) 2025-09-07T08:25:30.0040442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:25:30.0040780Z return forward_call(*args, **kwargs) 2025-09-07T08:25:30.0041123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 446, in forward 2025-09-07T08:25:30.0041504Z if max_word_piece_length + 2 > self._max_position_embeddings: 2025-09-07T08:25:30.0041661Z 2025-09-07T08:25:30.0041990Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-09-07T08:25:30.0042355Z 2025-09-07T08:25:30.0042358Z 2025-09-07T08:25:30.0042813Z The error above occurred when calling torch.export.export. If you would like to view some more information about this error, and get a list of all other errors that may occur in your export call, you can replace your `export()` call with `draft_export()`. 2025-09-07T08:25:30.0043446Z TorchDynamo optimized model failed to run because of following error 2025-09-07T08:25:30.1141417Z fail_to_run 2025-09-07T08:25:30.1141749Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:31.0921441Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:25:31.0923038Z import pynvml # type: ignore[import] 2025-09-07T08:25:32.8175242Z 2025-09-07T08:25:33.1194814Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:25:33.1195304Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:25:33.1195698Z cpu eval functorch_dp_cifar10 2025-09-07T08:25:33.1436742Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:33.1574072Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:33.1684197Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:40.5956352Z pass 2025-09-07T08:25:40.5957900Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:41.5450462Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:25:41.5451879Z import pynvml # type: ignore[import] 2025-09-07T08:25:43.2629571Z 2025-09-07T08:25:43.3875726Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:25:43.3876160Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:25:43.3876493Z cpu eval functorch_maml_omniglot 2025-09-07T08:25:43.3911643Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:43.3950604Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:43.3977654Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:50.5537687Z pass 2025-09-07T08:25:50.5538354Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:51.4226317Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:25:51.4227819Z import pynvml # type: ignore[import] 2025-09-07T08:25:53.1411369Z 2025-09-07T08:25:54.4313757Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:25:54.4314464Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:25:54.4314714Z cpu eval hf_Albert 2025-09-07T08:25:54.8925743Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:54.9768061Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:55.0592435Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:04.0258634Z pass 2025-09-07T08:26:04.0259103Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:05.0646321Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:26:05.0647765Z import pynvml # type: ignore[import] 2025-09-07T08:26:06.7840929Z 2025-09-07T08:26:10.0898209Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:26:10.0898616Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:26:10.0898938Z cpu eval hf_Bart 2025-09-07T08:26:10.7885180Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:10.9672078Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:11.1235049Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:20.8939550Z pass 2025-09-07T08:26:20.8939966Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:22.1319088Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:26:22.1320536Z import pynvml # type: ignore[import] 2025-09-07T08:26:23.8506431Z 2025-09-07T08:26:26.2221492Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:26:26.2221812Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:26:26.2222068Z cpu eval hf_Bert 2025-09-07T08:26:26.7378368Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:26.8491709Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:26.9590035Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:35.9722818Z pass 2025-09-07T08:26:35.9723252Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:37.1331913Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:26:37.1333380Z import pynvml # type: ignore[import] 2025-09-07T08:26:38.8542583Z 2025-09-07T08:26:42.7072068Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:26:42.7072394Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:26:42.7072635Z cpu eval hf_Bert_large 2025-09-07T08:26:43.9483535Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:44.2299715Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:44.4859881Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:55.4484266Z pass 2025-09-07T08:26:55.4484704Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:56.6933330Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:26:56.6934777Z import pynvml # type: ignore[import] 2025-09-07T08:26:58.4131612Z 2025-09-07T08:26:59.5611746Z loading model: 0it [00:00, ?it/s]BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:26:59.5612973Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:26:59.5613727Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:26:59.5614310Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:26:59.5615510Z WARNING:transformers.modeling_utils:BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:26:59.5617187Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:26:59.5618224Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:26:59.5619058Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:27:00.9151524Z Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:27:00.9152571Z WARNING:transformers.models.big_bird.modeling_big_bird:Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:27:01.9558395Z 2025-09-07T08:27:01.9558877Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:27:01.9559260Z cpu eval hf_BigBird 2025-09-07T08:27:03.3241653Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:04.7944478Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:06.2580157Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:07.3688691Z ERROR:common: 2025-09-07T08:27:07.3688938Z Traceback (most recent call last): 2025-09-07T08:27:07.3689317Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 2320, in check_accuracy 2025-09-07T08:27:07.3689659Z optimized_model_iter_fn = optimize_ctx( 2025-09-07T08:27:07.3690341Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 1523, in export 2025-09-07T08:27:07.3690690Z ep = torch.export.export( 2025-09-07T08:27:07.3691034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 311, in export 2025-09-07T08:27:07.3691416Z raise e 2025-09-07T08:27:07.3691702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 277, in export 2025-09-07T08:27:07.3692067Z return _export( 2025-09-07T08:27:07.3692360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:27:07.3692665Z raise e 2025-09-07T08:27:07.3692938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:27:07.3693244Z ep = fn(*args, **kwargs) 2025-09-07T08:27:07.3693568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:27:07.3693903Z return fn(*args, **kwargs) 2025-09-07T08:27:07.3694203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2255, in _export 2025-09-07T08:27:07.3694508Z ep = _export_for_training( 2025-09-07T08:27:07.3694798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:27:07.3695099Z raise e 2025-09-07T08:27:07.3695365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:27:07.3695669Z ep = fn(*args, **kwargs) 2025-09-07T08:27:07.3695990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:27:07.3696316Z return fn(*args, **kwargs) 2025-09-07T08:27:07.3696671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2071, in _export_for_training 2025-09-07T08:27:07.3697018Z export_artifact = export_func( 2025-09-07T08:27:07.3697356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1415, in _strict_export 2025-09-07T08:27:07.3697715Z gm_torch_level = _export_to_torch_ir( 2025-09-07T08:27:07.3698069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 812, in _export_to_torch_ir 2025-09-07T08:27:07.3698444Z gm_torch_level, _ = torch._dynamo.export( 2025-09-07T08:27:07.3698808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 2002, in inner 2025-09-07T08:27:07.3699168Z result_traced = opt_f(*args, **kwargs) 2025-09-07T08:27:07.3699514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 414, in __call__ 2025-09-07T08:27:07.3699859Z return super().__call__(*args, **kwargs) 2025-09-07T08:27:07.3700240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl 2025-09-07T08:27:07.3700632Z return self._call_impl(*args, **kwargs) 2025-09-07T08:27:07.3700984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:27:07.3701327Z return forward_call(*args, **kwargs) 2025-09-07T08:27:07.3701666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 841, in compile_wrapper 2025-09-07T08:27:07.3702065Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-09-07T08:27:07.3702494Z torch._dynamo.exc.UserError: Could not guard on data-dependent expression Eq(u1, u0) (unhinted: Eq(u1, u0)). (Size-like symbols: u1, u0) 2025-09-07T08:27:07.3702791Z 2025-09-07T08:27:07.3705608Z consider using data-dependent friendly APIs such as guard_or_false, guard_or_true and statically_known_trueCaused by: rand_attn = np.stack(rand_attn, axis=0) # transformers/models/big_bird/modeling_big_bird.py:574 in bigbird_block_sparse_attention (_refs/__init__.py:3992 in _check_stack_inputs) 2025-09-07T08:27:07.3706356Z For more information, run with TORCH_LOGS="dynamic" 2025-09-07T08:27:07.3706681Z For extended logs when we create symbols, also add TORCHDYNAMO_EXTENDED_DEBUG_CREATE_SYMBOL="u1,u0" 2025-09-07T08:27:07.3707072Z If you suspect the guard was triggered from C++, add TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:27:07.3707596Z For more debugging help, see https://docs.google.com/document/d/1HSuTTVvYH1pTew89Rtpeu84Ht3nQEFTYhAX3Ypa_xJs/edit?usp=sharing 2025-09-07T08:27:07.3707954Z 2025-09-07T08:27:07.3708029Z User Stack (most recent call last): 2025-09-07T08:27:07.3708237Z (snipped, see stack below for prefix) 2025-09-07T08:27:07.3708618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2303, in forward 2025-09-07T08:27:07.3708989Z outputs = self.bert( 2025-09-07T08:27:07.3709307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:27:07.3709640Z return forward_call(*args, **kwargs) 2025-09-07T08:27:07.3710015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1991, in forward 2025-09-07T08:27:07.3710386Z encoder_outputs = self.encoder( 2025-09-07T08:27:07.3710715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:27:07.3711044Z return forward_call(*args, **kwargs) 2025-09-07T08:27:07.3711416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1585, in forward 2025-09-07T08:27:07.3711780Z layer_outputs = layer_module( 2025-09-07T08:27:07.3712101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:27:07.3712429Z return forward_call(*args, **kwargs) 2025-09-07T08:27:07.3712801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1468, in forward 2025-09-07T08:27:07.3713183Z self_attention_outputs = self.attention( 2025-09-07T08:27:07.3713512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:27:07.3713843Z return forward_call(*args, **kwargs) 2025-09-07T08:27:07.3714216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1381, in forward 2025-09-07T08:27:07.3714586Z self_outputs = self.self( 2025-09-07T08:27:07.3714902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:27:07.3715223Z return forward_call(*args, **kwargs) 2025-09-07T08:27:07.3715590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 455, in forward 2025-09-07T08:27:07.3716022Z context_layer, attention_probs = self.bigbird_block_sparse_attention( 2025-09-07T08:27:07.3716501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 574, in bigbird_block_sparse_attention 2025-09-07T08:27:07.3716925Z rand_attn = np.stack(rand_attn, axis=0) 2025-09-07T08:27:07.3717053Z 2025-09-07T08:27:07.3717165Z For C++ stack trace, run with TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:27:07.3717597Z For more information about this error, see: https://pytorch.org/docs/main/generated/exportdb/index.html#constrain-as-size-example 2025-09-07T08:27:07.3717910Z 2025-09-07T08:27:07.3717972Z from user code: 2025-09-07T08:27:07.3718310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2303, in forward 2025-09-07T08:27:07.3718671Z outputs = self.bert( 2025-09-07T08:27:07.3719040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:27:07.3719368Z return forward_call(*args, **kwargs) 2025-09-07T08:27:07.3719744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1991, in forward 2025-09-07T08:27:07.3720115Z encoder_outputs = self.encoder( 2025-09-07T08:27:07.3720459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:27:07.3720795Z return forward_call(*args, **kwargs) 2025-09-07T08:27:07.3721166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1585, in forward 2025-09-07T08:27:07.3721533Z layer_outputs = layer_module( 2025-09-07T08:27:07.3721863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:27:07.3722190Z return forward_call(*args, **kwargs) 2025-09-07T08:27:07.3722555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1468, in forward 2025-09-07T08:27:07.3722941Z self_attention_outputs = self.attention( 2025-09-07T08:27:07.3723278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:27:07.3723607Z return forward_call(*args, **kwargs) 2025-09-07T08:27:07.3723971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1381, in forward 2025-09-07T08:27:07.3724338Z self_outputs = self.self( 2025-09-07T08:27:07.3724656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:27:07.3724984Z return forward_call(*args, **kwargs) 2025-09-07T08:27:07.3725352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 455, in forward 2025-09-07T08:27:07.3725777Z context_layer, attention_probs = self.bigbird_block_sparse_attention( 2025-09-07T08:27:07.3726252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 574, in bigbird_block_sparse_attention 2025-09-07T08:27:07.3726679Z rand_attn = np.stack(rand_attn, axis=0) 2025-09-07T08:27:07.3726805Z 2025-09-07T08:27:07.3727135Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-09-07T08:27:07.3727501Z 2025-09-07T08:27:07.3727504Z 2025-09-07T08:27:07.3727957Z The error above occurred when calling torch.export.export. If you would like to view some more information about this error, and get a list of all other errors that may occur in your export call, you can replace your `export()` call with `draft_export()`. 2025-09-07T08:27:07.3728589Z TorchDynamo optimized model failed to run because of following error 2025-09-07T08:27:07.5381569Z fail_to_run 2025-09-07T08:27:07.5381968Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:08.7820920Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:27:08.7822378Z import pynvml # type: ignore[import] 2025-09-07T08:27:10.4963092Z 2025-09-07T08:27:12.5506660Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:27:12.5506975Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:27:12.5507218Z cpu eval hf_DistilBert 2025-09-07T08:27:12.8665608Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:12.9307305Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:12.9929952Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:20.8056202Z pass 2025-09-07T08:27:20.8056626Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:21.8923046Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:27:21.8924860Z import pynvml # type: ignore[import] 2025-09-07T08:27:23.6101510Z 2025-09-07T08:27:27.3895106Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:27:27.3896193Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:27:27.3897060Z cpu eval hf_GPT2 2025-09-07T08:27:27.6350758Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:27.7113608Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:27.7837162Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:37.4109682Z pass 2025-09-07T08:27:37.4111558Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:38.6497173Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:27:38.6498592Z import pynvml # type: ignore[import] 2025-09-07T08:27:40.3765342Z 2025-09-07T08:27:51.0752991Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:27:51.0753319Z loading model: 0it [00:10, ?it/s] 2025-09-07T08:27:51.0753605Z cpu eval hf_GPT2_large 2025-09-07T08:27:51.4510957Z pass_due_to_skip 2025-09-07T08:27:51.4511318Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:53.0417703Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:27:53.0419154Z import pynvml # type: ignore[import] 2025-09-07T08:27:54.7579607Z 2025-09-07T08:27:57.8535604Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:27:57.8535954Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:27:57.8536644Z cpu eval hf_Longformer 2025-09-07T08:27:58.8674350Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:00.0118237Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:01.1511422Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:02.2611577Z ERROR:common: 2025-09-07T08:28:02.2611833Z Traceback (most recent call last): 2025-09-07T08:28:02.2612251Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 2320, in check_accuracy 2025-09-07T08:28:02.2612585Z optimized_model_iter_fn = optimize_ctx( 2025-09-07T08:28:02.2612909Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 1523, in export 2025-09-07T08:28:02.2613211Z ep = torch.export.export( 2025-09-07T08:28:02.2613547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 311, in export 2025-09-07T08:28:02.2613863Z raise e 2025-09-07T08:28:02.2614136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 277, in export 2025-09-07T08:28:02.2614443Z return _export( 2025-09-07T08:28:02.2615070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:28:02.2615385Z raise e 2025-09-07T08:28:02.2615655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:28:02.2616022Z ep = fn(*args, **kwargs) 2025-09-07T08:28:02.2616340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:28:02.2616728Z return fn(*args, **kwargs) 2025-09-07T08:28:02.2617035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2255, in _export 2025-09-07T08:28:02.2617354Z ep = _export_for_training( 2025-09-07T08:28:02.2617654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:28:02.2617946Z raise e 2025-09-07T08:28:02.2618213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:28:02.2618514Z ep = fn(*args, **kwargs) 2025-09-07T08:28:02.2618834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:28:02.2619194Z return fn(*args, **kwargs) 2025-09-07T08:28:02.2619514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2071, in _export_for_training 2025-09-07T08:28:02.2619859Z export_artifact = export_func( 2025-09-07T08:28:02.2620182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1415, in _strict_export 2025-09-07T08:28:02.2620512Z gm_torch_level = _export_to_torch_ir( 2025-09-07T08:28:02.2620873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 812, in _export_to_torch_ir 2025-09-07T08:28:02.2621242Z gm_torch_level, _ = torch._dynamo.export( 2025-09-07T08:28:02.2621602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 2002, in inner 2025-09-07T08:28:02.2621948Z result_traced = opt_f(*args, **kwargs) 2025-09-07T08:28:02.2622298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 414, in __call__ 2025-09-07T08:28:02.2622652Z return super().__call__(*args, **kwargs) 2025-09-07T08:28:02.2623049Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl 2025-09-07T08:28:02.2623423Z return self._call_impl(*args, **kwargs) 2025-09-07T08:28:02.2623782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:28:02.2624128Z return forward_call(*args, **kwargs) 2025-09-07T08:28:02.2624461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 841, in compile_wrapper 2025-09-07T08:28:02.2624853Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-09-07T08:28:02.2625383Z torch._dynamo.exc.UserError: Consider annotating your code using torch._check*(). Could not guard on data-dependent expression Eq(u0, 1) (unhinted: Eq(u0, 1)). (Size-like symbols: none) 2025-09-07T08:28:02.2625780Z 2025-09-07T08:28:02.2626286Z consider using data-dependent friendly APIs such as guard_or_false, guard_or_true and statically_known_trueCaused by: if is_global_attn: # transformers/models/longformer/modeling_longformer.py:554 in forward (_dynamo/variables/tensor.py:1435 in evaluate_expr) 2025-09-07T08:28:02.2626958Z For more information, run with TORCH_LOGS="dynamic" 2025-09-07T08:28:02.2627289Z For extended logs when we create symbols, also add TORCHDYNAMO_EXTENDED_DEBUG_CREATE_SYMBOL="u0" 2025-09-07T08:28:02.2627666Z If you suspect the guard was triggered from C++, add TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:28:02.2628223Z For more debugging help, see https://docs.google.com/document/d/1HSuTTVvYH1pTew89Rtpeu84Ht3nQEFTYhAX3Ypa_xJs/edit?usp=sharing 2025-09-07T08:28:02.2628554Z 2025-09-07T08:28:02.2628634Z User Stack (most recent call last): 2025-09-07T08:28:02.2628836Z (snipped, see stack below for prefix) 2025-09-07T08:28:02.2629240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1703, in forward 2025-09-07T08:28:02.2629653Z outputs = self.longformer( 2025-09-07T08:28:02.2629980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:28:02.2630350Z return forward_call(*args, **kwargs) 2025-09-07T08:28:02.2630751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1600, in forward 2025-09-07T08:28:02.2631157Z encoder_outputs = self.encoder( 2025-09-07T08:28:02.2631485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:28:02.2631825Z return forward_call(*args, **kwargs) 2025-09-07T08:28:02.2632219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in forward 2025-09-07T08:28:02.2632613Z layer_outputs = layer_module( 2025-09-07T08:28:02.2632938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:28:02.2633281Z return forward_call(*args, **kwargs) 2025-09-07T08:28:02.2633676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:28:02.2634079Z self_attn_outputs = self.attention( 2025-09-07T08:28:02.2634415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:28:02.2634738Z return forward_call(*args, **kwargs) 2025-09-07T08:28:02.2635135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:28:02.2635524Z self_outputs = self.self( 2025-09-07T08:28:02.2635842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:28:02.2636168Z return forward_call(*args, **kwargs) 2025-09-07T08:28:02.2636563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 554, in forward 2025-09-07T08:28:02.2636947Z if is_global_attn: 2025-09-07T08:28:02.2637049Z 2025-09-07T08:28:02.2637175Z For C++ stack trace, run with TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:28:02.2637608Z For more information about this error, see: https://pytorch.org/docs/main/generated/exportdb/index.html#constrain-as-size-example 2025-09-07T08:28:02.2637920Z 2025-09-07T08:28:02.2637980Z from user code: 2025-09-07T08:28:02.2638349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1703, in forward 2025-09-07T08:28:02.2638745Z outputs = self.longformer( 2025-09-07T08:28:02.2639071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:28:02.2639413Z return forward_call(*args, **kwargs) 2025-09-07T08:28:02.2639801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1600, in forward 2025-09-07T08:28:02.2640197Z encoder_outputs = self.encoder( 2025-09-07T08:28:02.2640530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:28:02.2640863Z return forward_call(*args, **kwargs) 2025-09-07T08:28:02.2641245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in forward 2025-09-07T08:28:02.2641640Z layer_outputs = layer_module( 2025-09-07T08:28:02.2642005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:28:02.2642339Z return forward_call(*args, **kwargs) 2025-09-07T08:28:02.2642731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:28:02.2643135Z self_attn_outputs = self.attention( 2025-09-07T08:28:02.2643466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:28:02.2643813Z return forward_call(*args, **kwargs) 2025-09-07T08:28:02.2644201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:28:02.2644610Z self_outputs = self.self( 2025-09-07T08:28:02.2644930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:28:02.2645258Z return forward_call(*args, **kwargs) 2025-09-07T08:28:02.2645642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 554, in forward 2025-09-07T08:28:02.2646019Z if is_global_attn: 2025-09-07T08:28:02.2646126Z 2025-09-07T08:28:02.2646445Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-09-07T08:28:02.2646820Z 2025-09-07T08:28:02.2646823Z 2025-09-07T08:28:02.2647276Z The error above occurred when calling torch.export.export. If you would like to view some more information about this error, and get a list of all other errors that may occur in your export call, you can replace your `export()` call with `draft_export()`. 2025-09-07T08:28:02.2647911Z TorchDynamo optimized model failed to run because of following error 2025-09-07T08:28:02.4006694Z fail_to_run 2025-09-07T08:28:02.4007125Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:03.5825649Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:28:03.5827073Z import pynvml # type: ignore[import] 2025-09-07T08:28:05.2982979Z 2025-09-07T08:28:06.3572822Z loading model: 0it [00:00, ?it/s]Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T08:28:06.3575212Z WARNING:transformers.models.reformer.modeling_reformer:Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T08:28:06.7580269Z 2025-09-07T08:28:06.7580868Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:28:06.7581354Z cpu eval hf_Reformer 2025-09-07T08:28:07.1004230Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:07.8669087Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:08.6290257Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:09.0903228Z class GraphModule(torch.nn.Module): 2025-09-07T08:28:09.0903743Z def forward(self, L_input_ids_: "i64[4, 2048][2048, 1]cpu"): 2025-09-07T08:28:09.0904147Z l_input_ids_ = L_input_ids_ 2025-09-07T08:28:09.0904437Z 2025-09-07T08:28:09.0905852Z # File: /opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py:194 in _get_least_common_mult_chunk_len, code: return np.lcm(config.lsh_attn_chunk_length, config.local_attn_chunk_length) 2025-09-07T08:28:09.0907114Z least_common_mult_chunk_length: "i64[][]cpu" = torch__dynamo_utils_wrapped_lcm(64, 64) 2025-09-07T08:28:09.0907616Z 2025-09-07T08:28:09.0908412Z # File: /opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py:2099 in forward, code: input_shape[-1] % least_common_mult_chunk_length != 0 2025-09-07T08:28:09.0909651Z wrapped_mod: "i64[][]cpu" = torch__dynamo_utils_wrapped_mod(2048, least_common_mult_chunk_length); least_common_mult_chunk_length = None 2025-09-07T08:28:09.0910462Z wrapped_ne: "b8[][]cpu" = torch__dynamo_utils_wrapped_ne(wrapped_mod, 0); wrapped_mod = wrapped_ne = None 2025-09-07T08:28:09.0910925Z 2025-09-07T08:28:09.0911172Z class GraphModule(torch.nn.Module): 2025-09-07T08:28:09.0911573Z def forward(self, L_input_ids_: "i64[4, 2048][2048, 1]cpu"): 2025-09-07T08:28:09.0911961Z l_input_ids_ = L_input_ids_ 2025-09-07T08:28:09.0912238Z 2025-09-07T08:28:09.0913210Z # File: /opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py:194 in _get_least_common_mult_chunk_len, code: return np.lcm(config.lsh_attn_chunk_length, config.local_attn_chunk_length) 2025-09-07T08:28:09.0914449Z least_common_mult_chunk_length: "i64[][]cpu" = torch__dynamo_utils_wrapped_lcm(64, 64) 2025-09-07T08:28:09.0914910Z 2025-09-07T08:28:09.0915740Z # File: /opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py:2099 in forward, code: input_shape[-1] % least_common_mult_chunk_length != 0 2025-09-07T08:28:09.0916941Z wrapped_mod: "i64[][]cpu" = torch__dynamo_utils_wrapped_mod(2048, least_common_mult_chunk_length); least_common_mult_chunk_length = None 2025-09-07T08:28:09.0917801Z wrapped_ne: "b8[][]cpu" = torch__dynamo_utils_wrapped_ne(wrapped_mod, 0); wrapped_mod = wrapped_ne = None 2025-09-07T08:28:09.0918325Z 2025-09-07T08:28:09.0918528Z ERROR:common: 2025-09-07T08:28:09.0918759Z Traceback (most recent call last): 2025-09-07T08:28:09.0919295Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 2320, in check_accuracy 2025-09-07T08:28:09.0919868Z optimized_model_iter_fn = optimize_ctx( 2025-09-07T08:28:09.0920410Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 1523, in export 2025-09-07T08:28:09.0920920Z ep = torch.export.export( 2025-09-07T08:28:09.0921474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 311, in export 2025-09-07T08:28:09.0921982Z raise e 2025-09-07T08:28:09.0922437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 277, in export 2025-09-07T08:28:09.0922959Z return _export( 2025-09-07T08:28:09.0923426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:28:09.0923937Z raise e 2025-09-07T08:28:09.0924397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:28:09.0924933Z ep = fn(*args, **kwargs) 2025-09-07T08:28:09.0925494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:28:09.0926080Z return fn(*args, **kwargs) 2025-09-07T08:28:09.0926613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2255, in _export 2025-09-07T08:28:09.0927160Z ep = _export_for_training( 2025-09-07T08:28:09.0927684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:28:09.0928204Z raise e 2025-09-07T08:28:09.0928657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:28:09.0929327Z ep = fn(*args, **kwargs) 2025-09-07T08:28:09.0929898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:28:09.0930482Z return fn(*args, **kwargs) 2025-09-07T08:28:09.0931047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2071, in _export_for_training 2025-09-07T08:28:09.0931685Z export_artifact = export_func( 2025-09-07T08:28:09.0932296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1415, in _strict_export 2025-09-07T08:28:09.0932869Z gm_torch_level = _export_to_torch_ir( 2025-09-07T08:28:09.0933468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 812, in _export_to_torch_ir 2025-09-07T08:28:09.0934067Z gm_torch_level, _ = torch._dynamo.export( 2025-09-07T08:28:09.0934645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 2002, in inner 2025-09-07T08:28:09.0935209Z result_traced = opt_f(*args, **kwargs) 2025-09-07T08:28:09.0935776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 414, in __call__ 2025-09-07T08:28:09.0936334Z return super().__call__(*args, **kwargs) 2025-09-07T08:28:09.0936964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl 2025-09-07T08:28:09.0937595Z return self._call_impl(*args, **kwargs) 2025-09-07T08:28:09.0938176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:28:09.0938748Z return forward_call(*args, **kwargs) 2025-09-07T08:28:09.0939321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 841, in compile_wrapper 2025-09-07T08:28:09.0940014Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-09-07T08:28:09.0940510Z torch._dynamo.exc.Unsupported: Data-dependent branching 2025-09-07T08:28:09.0941223Z Explanation: Detected data-dependent branching (e.g. `if my_tensor.sum() > 0:`). Dynamo does not support tracing dynamic control flow. 2025-09-07T08:28:09.0942187Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-09-07T08:28:09.0942904Z Hint: Use `torch.cond` to express dynamic control flow. 2025-09-07T08:28:09.0943180Z 2025-09-07T08:28:09.0943414Z Developer debug context: attempted to jump with NumpyNdarrayVariable() 2025-09-07T08:28:09.0943767Z 2025-09-07T08:28:09.0944221Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0170.html 2025-09-07T08:28:09.0944752Z 2025-09-07T08:28:09.0944848Z from user code: 2025-09-07T08:28:09.0945489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2481, in forward 2025-09-07T08:28:09.0946175Z reformer_outputs = self.reformer( 2025-09-07T08:28:09.0946766Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:28:09.0947352Z return forward_call(*args, **kwargs) 2025-09-07T08:28:09.0948047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2099, in forward 2025-09-07T08:28:09.0948779Z input_shape[-1] % least_common_mult_chunk_length != 0 2025-09-07T08:28:09.0949039Z 2025-09-07T08:28:09.0949618Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-09-07T08:28:09.0950296Z 2025-09-07T08:28:09.0950512Z TorchDynamo optimized model failed to run because of following error 2025-09-07T08:28:09.1549223Z fail_to_run 2025-09-07T08:28:09.1550076Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:10.1308447Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:28:10.1310224Z import pynvml # type: ignore[import] 2025-09-07T08:28:11.8389025Z 2025-09-07T08:28:16.6276692Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:28:16.6277011Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:28:16.6277245Z cpu eval hf_Roberta_base 2025-09-07T08:28:17.9681460Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:18.1983405Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:18.4228464Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:29.1940339Z pass 2025-09-07T08:28:29.1940759Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:30.5832927Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:28:30.5834538Z import pynvml # type: ignore[import] 2025-09-07T08:28:32.3092471Z 2025-09-07T08:28:34.4881641Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:28:34.4883124Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:28:34.4883504Z cpu eval hf_T5 2025-09-07T08:28:34.9477444Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:35.1825439Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:35.4138224Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:45.5863059Z pass 2025-09-07T08:28:45.5863477Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:46.8333047Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:28:46.8334530Z import pynvml # type: ignore[import] 2025-09-07T08:28:48.5540556Z 2025-09-07T08:28:58.7290162Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:28:58.7291172Z loading model: 0it [00:10, ?it/s] 2025-09-07T08:28:58.7292217Z cpu eval hf_T5_base 2025-09-07T08:29:11.6807105Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:29:19.5853806Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:29:27.4625957Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:29:56.9139026Z pass 2025-09-07T08:29:56.9139458Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:29:59.4356119Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:29:59.4357487Z import pynvml # type: ignore[import] 2025-09-07T08:30:01.1494372Z 2025-09-07T08:30:10.6434530Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:30:10.6434833Z loading model: 0it [00:09, ?it/s] 2025-09-07T08:30:10.6435583Z cpu eval hf_T5_large 2025-09-07T08:30:11.0351450Z pass_due_to_skip 2025-09-07T08:30:11.0351803Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:12.4705881Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:30:12.4707824Z import pynvml # type: ignore[import] 2025-09-07T08:30:14.1877449Z 2025-09-07T08:30:18.4727987Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:30:18.4728427Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:30:18.4728738Z cpu eval hf_distil_whisper 2025-09-07T08:30:22.7939974Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:23.5458107Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:24.3169081Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:36.4433765Z pass 2025-09-07T08:30:36.4434193Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:37.8390476Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:30:37.8391932Z import pynvml # type: ignore[import] 2025-09-07T08:30:39.5572679Z 2025-09-07T08:30:39.6477157Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:30:39.6477548Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:30:39.6477863Z cpu eval lennard_jones 2025-09-07T08:30:39.6486952Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:39.6505592Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:39.6514170Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:46.7455753Z pass 2025-09-07T08:30:46.7456230Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:47.2528964Z accuracy pass_rate=65.22% 2025-09-07T08:30:47.2529278Z calls_captured gmean=0.00x mean=260.870x 2025-09-07T08:30:47.2529518Z unique_graphs gmean=0.00x mean=0.652x 2025-09-07T08:30:47.2531546Z graph_breaks gmean=0.00x mean=0.000x 2025-09-07T08:30:47.2532257Z unique_graph_breaks gmean=0.00x mean=0.000x 2025-09-07T08:30:47.2534086Z autograd_captures gmean=0.00x mean=0.000x 2025-09-07T08:30:47.2535896Z autograd_compiles gmean=0.00x mean=0.000x 2025-09-07T08:30:47.2537724Z cudagraph_skips gmean=0.00x mean=0.000x 2025-09-07T08:30:47.2538466Z compilation_latency mean=5.775 seconds 2025-09-07T08:30:47.6385549Z + taskset -c 0-93 python benchmarks/dynamo/torchbench.py --accuracy --no-translation-validation --freezing --inference --amp --export-aot-inductor --disable-cudagraphs --device cpu --total-partitions 4 --partition-id 1 --output /var/lib/jenkins/workspace/test/test-reports/inductor_aot_inductor_torchbench_amp_inference_cpu_x86_accuracy.csv 2025-09-07T08:30:47.9850450Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:30:47.9851867Z import pynvml # type: ignore[import] 2025-09-07T08:30:50.0521734Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:30:50.0523202Z import pynvml # type: ignore[import] 2025-09-07T08:30:51.7783112Z 2025-09-07T08:30:56.8742737Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:30:56.8745103Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:30:56.8747991Z cpu eval dlrm 2025-09-07T08:30:57.0696423Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:57.1189255Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:57.1669331Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:19.3375885Z pass 2025-09-07T08:31:19.3376471Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:20.8314036Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:31:20.8315429Z import pynvml # type: ignore[import] 2025-09-07T08:31:22.5473918Z 2025-09-07T08:31:22.8486677Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:31:22.8487083Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:31:22.8489803Z cpu eval functorch_dp_cifar10 2025-09-07T08:31:22.8714226Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:22.8849478Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:22.8959084Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:38.0527609Z pass 2025-09-07T08:31:38.0528011Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:39.3836108Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:31:39.3837569Z import pynvml # type: ignore[import] 2025-09-07T08:31:41.1005216Z 2025-09-07T08:31:41.2286010Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:31:41.2286390Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:31:41.2286726Z cpu eval functorch_maml_omniglot 2025-09-07T08:31:41.2321475Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:41.2360106Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:41.2385892Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:51.1101790Z pass 2025-09-07T08:31:51.1102259Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:52.2208594Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:31:52.2210042Z import pynvml # type: ignore[import] 2025-09-07T08:31:53.9308377Z 2025-09-07T08:31:55.2117842Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:31:55.2118161Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:31:55.2118412Z cpu eval hf_Albert 2025-09-07T08:31:55.6763347Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:55.7612005Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:55.8440647Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:12.0072682Z pass 2025-09-07T08:32:12.0074522Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:13.5351900Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:32:13.5353321Z import pynvml # type: ignore[import] 2025-09-07T08:32:15.2626504Z 2025-09-07T08:32:18.5897339Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:32:18.5897755Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:32:18.5898125Z cpu eval hf_Bart 2025-09-07T08:32:19.2723871Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:19.4506512Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:19.6046767Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:41.3501657Z pass 2025-09-07T08:32:41.3502095Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:43.2720261Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:32:43.2721691Z import pynvml # type: ignore[import] 2025-09-07T08:32:44.9859214Z 2025-09-07T08:32:47.3663124Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:32:47.3663435Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:32:47.3663688Z cpu eval hf_Bert 2025-09-07T08:32:47.9013201Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:48.0152847Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:48.1259825Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:05.8714848Z pass 2025-09-07T08:33:05.8715250Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:07.5217732Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:33:07.5219192Z import pynvml # type: ignore[import] 2025-09-07T08:33:09.2328734Z 2025-09-07T08:33:13.0976089Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:33:13.0976427Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:33:13.0976749Z cpu eval hf_Bert_large 2025-09-07T08:33:14.3421604Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:14.5986558Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:14.8554681Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:41.0123459Z pass 2025-09-07T08:33:41.0123881Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:43.0171226Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:33:43.0172771Z import pynvml # type: ignore[import] 2025-09-07T08:33:44.7297950Z 2025-09-07T08:33:45.8755637Z loading model: 0it [00:00, ?it/s]BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:33:45.8757243Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:33:45.8758005Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:33:45.8758589Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:33:45.8759755Z WARNING:transformers.modeling_utils:BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:33:45.8761013Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:33:45.8761858Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:33:45.8762713Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:33:47.2289440Z Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:33:47.2290744Z WARNING:transformers.models.big_bird.modeling_big_bird:Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:33:48.2754743Z 2025-09-07T08:33:48.2755252Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:33:48.2755626Z cpu eval hf_BigBird 2025-09-07T08:33:49.6431817Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:51.1214994Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:52.5873586Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:35:02.4367459Z pass 2025-09-07T08:35:02.4367981Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:35:05.4332783Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:35:05.4334258Z import pynvml # type: ignore[import] 2025-09-07T08:35:07.1454996Z 2025-09-07T08:35:09.2118379Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:35:09.2118713Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:35:09.2118961Z cpu eval hf_DistilBert 2025-09-07T08:35:09.5256849Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:35:09.5898166Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:35:09.6524217Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:35:28.7495318Z pass 2025-09-07T08:35:28.7495749Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:35:30.2900027Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:35:30.2901902Z import pynvml # type: ignore[import] 2025-09-07T08:35:32.0219540Z 2025-09-07T08:35:35.7928832Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:35:35.7929386Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:35:35.7930237Z cpu eval hf_GPT2 2025-09-07T08:35:36.0392697Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:35:36.1149290Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:35:36.1882095Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:35:54.9009231Z pass 2025-09-07T08:35:54.9009710Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:35:56.6625809Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:35:56.6627215Z import pynvml # type: ignore[import] 2025-09-07T08:35:58.3763863Z 2025-09-07T08:36:09.1375259Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:36:09.1375608Z loading model: 0it [00:10, ?it/s] 2025-09-07T08:36:09.1375860Z cpu eval hf_GPT2_large 2025-09-07T08:36:09.5140856Z pass_due_to_skip 2025-09-07T08:36:09.5141236Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:36:11.0724986Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:36:11.0726440Z import pynvml # type: ignore[import] 2025-09-07T08:36:12.7796160Z 2025-09-07T08:36:17.5708754Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:36:17.5709056Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:36:17.5709301Z cpu eval hf_Roberta_base 2025-09-07T08:36:18.8948528Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:36:19.1203916Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:36:19.3458473Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:36:45.9365790Z pass 2025-09-07T08:36:45.9366270Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:36:48.0539775Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:36:48.0541222Z import pynvml # type: ignore[import] 2025-09-07T08:36:49.7604169Z 2025-09-07T08:36:51.9564176Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:36:51.9565190Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:36:51.9566181Z cpu eval hf_T5 2025-09-07T08:36:52.4149606Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:36:52.6460603Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:36:52.8789267Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:13.9556496Z pass 2025-09-07T08:37:13.9557302Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:15.7231019Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:37:15.7232422Z import pynvml # type: ignore[import] 2025-09-07T08:37:17.4337358Z 2025-09-07T08:37:27.4589519Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:37:27.4591003Z loading model: 0it [00:10, ?it/s] 2025-09-07T08:37:27.4591964Z cpu eval hf_T5_base 2025-09-07T08:37:40.4278296Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:48.2945107Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:56.1657168Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:46.6294989Z pass 2025-09-07T08:38:46.6295461Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:50.1377536Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:38:50.1378953Z import pynvml # type: ignore[import] 2025-09-07T08:38:51.8504202Z 2025-09-07T08:39:01.3763674Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:39:01.3764080Z loading model: 0it [00:09, ?it/s] 2025-09-07T08:39:01.3764394Z cpu eval hf_T5_large 2025-09-07T08:39:01.7671158Z pass_due_to_skip 2025-09-07T08:39:01.7672196Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:03.2829396Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:39:03.2830793Z import pynvml # type: ignore[import] 2025-09-07T08:39:04.9932098Z 2025-09-07T08:39:09.2716236Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:39:09.2716595Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:39:09.2716972Z cpu eval hf_distil_whisper 2025-09-07T08:39:13.5728045Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:14.3283486Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:15.1034764Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:43.1791997Z pass 2025-09-07T08:39:43.1792481Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:45.2949667Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:39:45.2951088Z import pynvml # type: ignore[import] 2025-09-07T08:39:47.0066736Z 2025-09-07T08:39:47.1029169Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:39:47.1029568Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:39:47.1029891Z cpu eval lennard_jones 2025-09-07T08:39:47.1038942Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:47.1058067Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:47.1066695Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:56.0700660Z pass 2025-09-07T08:39:56.0701137Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:56.7943585Z accuracy pass_rate=88.24% 2025-09-07T08:39:56.7947622Z calls_captured gmean=0.00x mean=0.000x 2025-09-07T08:39:56.7950066Z unique_graphs gmean=0.00x mean=0.000x 2025-09-07T08:39:56.7952603Z graph_breaks gmean=0.00x mean=0.000x 2025-09-07T08:39:56.7955129Z unique_graph_breaks gmean=0.00x mean=0.000x 2025-09-07T08:39:56.7957488Z autograd_captures gmean=0.00x mean=0.000x 2025-09-07T08:39:56.7959765Z autograd_compiles gmean=0.00x mean=0.000x 2025-09-07T08:39:56.7962353Z cudagraph_skips gmean=0.00x mean=0.000x 2025-09-07T08:39:56.7963007Z compilation_latency mean=0.000 seconds 2025-09-07T08:39:57.1874375Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true == *maxautotune-true* ]] 2025-09-07T08:39:57.1875276Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true == *cudagraphs_low_precision-true* ]] 2025-09-07T08:39:57.1875799Z + for target in "${targets[@]}" 2025-09-07T08:39:57.1875996Z + target_flag=('--performance') 2025-09-07T08:39:57.1876187Z + local target_flag 2025-09-07T08:39:57.1876371Z + [[ performance == \p\e\r\f\o\r\m\a\n\c\e ]] 2025-09-07T08:39:57.1876606Z + target_flag+=(--cold-start-latency) 2025-09-07T08:39:57.1877084Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true == *freezing-true* ]] 2025-09-07T08:39:57.1877534Z + target_flag+=(--freezing) 2025-09-07T08:39:57.1877984Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true-freezing-true == *default-true* ]] 2025-09-07T08:39:57.1879116Z + taskset -c 0-93 python benchmarks/dynamo/torchbench.py --performance --cold-start-latency --freezing --inference --amp --backend inductor --disable-cudagraphs --device cpu --total-partitions 4 --partition-id 1 --output /var/lib/jenkins/workspace/test/test-reports/inductor_no_cudagraphs_torchbench_amp_inference_cpu_x86_performance.csv 2025-09-07T08:39:57.5119308Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:39:57.5120725Z import pynvml # type: ignore[import] 2025-09-07T08:39:59.5726221Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:39:59.5727709Z import pynvml # type: ignore[import] 2025-09-07T08:40:01.2859803Z 2025-09-07T08:40:06.7450622Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:40:06.7451491Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:40:06.7452088Z cpu eval dlrm 2025-09-07T08:40:23.1807317Z 2025-09-07T08:40:23.2884803Z running benchmark: 0% 0/30 [00:00 2025-09-07T08:54:49.4780935Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4781126Z 2025-09-07T08:54:49.4781223Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4781677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4782098Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4782405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4782714Z x = module(x) 2025-09-07T08:54:49.4782998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4783313Z out = self.relu(out) 2025-09-07T08:54:49.4783408Z 2025-09-07T08:54:49.4783512Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4783942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4784350Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4784748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:54:49.4785210Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4785683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:54:49.4786142Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4786329Z 2025-09-07T08:54:49.4786427Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4786862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4787268Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4787587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4787891Z x = module(x) 2025-09-07T08:54:49.4788181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4788501Z out = self.relu(out) 2025-09-07T08:54:49.4788597Z 2025-09-07T08:54:49.4788703Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4789128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4789535Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4789849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4790163Z x = module(x) 2025-09-07T08:54:49.4790452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4790759Z out = self.relu(out) 2025-09-07T08:54:49.4790864Z 2025-09-07T08:54:49.4790963Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4791394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4791799Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4792102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4792414Z x = module(x) 2025-09-07T08:54:49.4792698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4793015Z out = self.relu(out) 2025-09-07T08:54:49.4793111Z 2025-09-07T08:54:49.4793228Z cudagraph partition due to non gpu ops 2025-09-07T08:54:49.4793447Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4793879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4794304Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4794621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4794936Z x = module(x) 2025-09-07T08:54:49.4795227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4795540Z out = self.relu(out) 2025-09-07T08:54:49.4795635Z 2025-09-07T08:54:49.4795739Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4796168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4796567Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4796879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4797186Z x = module(x) 2025-09-07T08:54:49.4797476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4797794Z out = self.relu(out) 2025-09-07T08:54:49.4797889Z 2025-09-07T08:54:49.4797985Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4798417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4798820Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4799134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4799432Z x = module(x) 2025-09-07T08:54:49.4799724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4800038Z out = self.relu(out) 2025-09-07T08:54:49.4800132Z 2025-09-07T08:54:49.4800237Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4800669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4801066Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4801382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4801693Z x = module(x) 2025-09-07T08:54:49.4801981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4802291Z out = self.relu(out) 2025-09-07T08:54:49.4802396Z 2025-09-07T08:54:49.4802469Z cudagraph partition due to non gpu ops 2025-09-07T08:54:49.4802698Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4803133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4803542Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4803850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4804170Z x = module(x) 2025-09-07T08:54:49.4804461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4804779Z out = self.relu(out) 2025-09-07T08:54:49.4804874Z 2025-09-07T08:54:49.4804977Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4805405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4805805Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4806831Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4807146Z x = module(x) 2025-09-07T08:54:49.4807427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4807773Z out = self.relu(out) 2025-09-07T08:54:49.4807880Z 2025-09-07T08:54:49.4807977Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4808431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4808842Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4809159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4809476Z x = module(x) 2025-09-07T08:54:49.4809769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4810092Z out = self.relu(out) 2025-09-07T08:54:49.4810194Z 2025-09-07T08:54:49.4810292Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4810729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4811140Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4811462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4811773Z x = module(x) 2025-09-07T08:54:49.4812059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4812382Z out = self.relu(out) 2025-09-07T08:54:49.4812488Z 2025-09-07T08:54:49.4812566Z cudagraph partition due to non gpu ops 2025-09-07T08:54:49.4812797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4813225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4813630Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4813948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4814261Z x = module(x) 2025-09-07T08:54:49.4814551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4814869Z out = self.relu(out) 2025-09-07T08:54:49.4814977Z 2025-09-07T08:54:49.4815075Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4815517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4815924Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4816245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4816553Z x = module(x) 2025-09-07T08:54:49.4816846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4817165Z out = self.relu(out) 2025-09-07T08:54:49.4817263Z 2025-09-07T08:54:49.4817372Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4817802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4818211Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4818535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4818848Z x = module(x) 2025-09-07T08:54:49.4819144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4819459Z out = self.relu(out) 2025-09-07T08:54:49.4819568Z 2025-09-07T08:54:49.4819699Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4820127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4820529Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4820857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4821172Z x = module(x) 2025-09-07T08:54:49.4821460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4821776Z out = self.relu(out) 2025-09-07T08:54:49.4821872Z 2025-09-07T08:54:49.4821975Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4822402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4822808Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4823119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4823424Z x = module(x) 2025-09-07T08:54:49.4823711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:54:49.4824018Z out = self.relu(out) 2025-09-07T08:54:49.4824124Z 2025-09-07T08:54:49.4824221Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4824653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4825058Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4825364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4825672Z x = module(x) 2025-09-07T08:54:49.4825962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:54:49.4826277Z out = self.relu(out) 2025-09-07T08:54:49.4826374Z 2025-09-07T08:54:49.4826455Z cudagraph partition due to non gpu ops 2025-09-07T08:54:49.4826672Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4827107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4827511Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4827907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:54:49.4828369Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4828836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:54:49.4829311Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4829498Z 2025-09-07T08:54:49.4829593Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4830021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4830429Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4830739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4831045Z x = module(x) 2025-09-07T08:54:49.4831333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4831646Z out = self.relu(out) 2025-09-07T08:54:49.4831744Z 2025-09-07T08:54:49.4831840Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4832299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4832707Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4833107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:54:49.4833587Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4834056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:54:49.4834555Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4834744Z 2025-09-07T08:54:49.4834839Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4835274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4835679Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4835998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4836306Z x = module(x) 2025-09-07T08:54:49.4836593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4836911Z out = self.relu(out) 2025-09-07T08:54:49.4837008Z 2025-09-07T08:54:49.4837112Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4837535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4837941Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4838255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4838561Z x = module(x) 2025-09-07T08:54:49.4838841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4839155Z out = self.relu(out) 2025-09-07T08:54:49.4839259Z 2025-09-07T08:54:49.4839354Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4839784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4840186Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4840493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4840801Z x = module(x) 2025-09-07T08:54:49.4841084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4841400Z out = self.relu(out) 2025-09-07T08:54:49.4841498Z 2025-09-07T08:54:49.4841582Z cudagraph partition due to non gpu ops 2025-09-07T08:54:49.4841801Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4842233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4842633Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4842945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4843248Z x = module(x) 2025-09-07T08:54:49.4843540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4843853Z out = self.relu(out) 2025-09-07T08:54:49.4843950Z 2025-09-07T08:54:49.4844056Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4844484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4844877Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4845221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4845532Z x = module(x) 2025-09-07T08:54:49.4845817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4846142Z out = self.relu(out) 2025-09-07T08:54:49.4846247Z 2025-09-07T08:54:49.4846342Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4846802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4847209Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4847518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4847820Z x = module(x) 2025-09-07T08:54:49.4848106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4848423Z out = self.relu(out) 2025-09-07T08:54:49.4848520Z 2025-09-07T08:54:49.4848624Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4849057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4849451Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4849766Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4850075Z x = module(x) 2025-09-07T08:54:49.4850364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4850673Z out = self.relu(out) 2025-09-07T08:54:49.4850776Z 2025-09-07T08:54:49.4850852Z cudagraph partition due to non gpu ops 2025-09-07T08:54:49.4851075Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4851505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4851916Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4852221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4852525Z x = module(x) 2025-09-07T08:54:49.4852809Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4853128Z out = self.relu(out) 2025-09-07T08:54:49.4853223Z 2025-09-07T08:54:49.4853319Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4853748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4854150Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4854468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4854771Z x = module(x) 2025-09-07T08:54:49.4855048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4855362Z out = self.relu(out) 2025-09-07T08:54:49.4855466Z 2025-09-07T08:54:49.4855564Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4855993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4856403Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4856705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4857010Z x = module(x) 2025-09-07T08:54:49.4857295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4857606Z out = self.relu(out) 2025-09-07T08:54:49.4857701Z 2025-09-07T08:54:49.4857824Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4858259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4858663Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4858996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4859324Z x = module(x) 2025-09-07T08:54:49.4859604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4859921Z out = self.relu(out) 2025-09-07T08:54:49.4860025Z 2025-09-07T08:54:49.4860101Z cudagraph partition due to non gpu ops 2025-09-07T08:54:49.4860327Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4860755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4861158Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4861472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4861779Z x = module(x) 2025-09-07T08:54:49.4862067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4862377Z out = self.relu(out) 2025-09-07T08:54:49.4862481Z 2025-09-07T08:54:49.4862578Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4863007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4863412Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4863717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4864026Z x = module(x) 2025-09-07T08:54:49.4864316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4864636Z out = self.relu(out) 2025-09-07T08:54:49.4864732Z 2025-09-07T08:54:49.4864836Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4865269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4865681Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4865999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4866312Z x = module(x) 2025-09-07T08:54:49.4866602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4866912Z out = self.relu(out) 2025-09-07T08:54:49.4867013Z 2025-09-07T08:54:49.4867108Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4867543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4867948Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4868303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4868614Z x = module(x) 2025-09-07T08:54:49.4868901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4869223Z out = self.relu(out) 2025-09-07T08:54:49.4869323Z 2025-09-07T08:54:49.4869406Z cudagraph partition due to non gpu ops 2025-09-07T08:54:49.4869624Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4870064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4870527Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4870843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4871140Z x = module(x) 2025-09-07T08:54:49.4871428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4871769Z out = self.relu(out) 2025-09-07T08:54:49.4871865Z 2025-09-07T08:54:49.4871971Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4872432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4872830Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4873151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4873463Z x = module(x) 2025-09-07T08:54:49.4873751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4874065Z out = self.relu(out) 2025-09-07T08:54:49.4874159Z 2025-09-07T08:54:49.4874254Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4874686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4875093Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4875406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4875703Z x = module(x) 2025-09-07T08:54:49.4875985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4876298Z out = self.relu(out) 2025-09-07T08:54:49.4876395Z 2025-09-07T08:54:49.4876498Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4876937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4877338Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4877653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4877970Z x = module(x) 2025-09-07T08:54:49.4878262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4878574Z out = self.relu(out) 2025-09-07T08:54:49.4878677Z 2025-09-07T08:54:49.4878753Z cudagraph partition due to non gpu ops 2025-09-07T08:54:49.4878978Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4879420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4879835Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4880150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4880463Z x = module(x) 2025-09-07T08:54:49.4880755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4881075Z out = self.relu(out) 2025-09-07T08:54:49.4881173Z 2025-09-07T08:54:49.4881276Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4881707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4882120Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4882431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4882745Z x = module(x) 2025-09-07T08:54:49.4883025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4883371Z out = self.relu(out) 2025-09-07T08:54:49.4883476Z 2025-09-07T08:54:49.4883574Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4884001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4884422Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4884725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4885052Z x = module(x) 2025-09-07T08:54:49.4885343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4885657Z out = self.relu(out) 2025-09-07T08:54:49.4885753Z 2025-09-07T08:54:49.4885850Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4886284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4886683Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4886992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4887296Z x = module(x) 2025-09-07T08:54:49.4887571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4887890Z out = self.relu(out) 2025-09-07T08:54:49.4887997Z 2025-09-07T08:54:49.4888093Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4888522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4888925Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4889230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4889534Z x = module(x) 2025-09-07T08:54:49.4889822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:54:49.4890137Z out = self.relu(out) 2025-09-07T08:54:49.4890233Z 2025-09-07T08:54:49.4890330Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4890768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4891171Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4891483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4891789Z x = module(x) 2025-09-07T08:54:49.4892068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:54:49.4892386Z out = self.relu(out) 2025-09-07T08:54:49.4892492Z 2025-09-07T08:54:49.4892569Z cudagraph partition due to non gpu ops 2025-09-07T08:54:49.4892797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4893224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4893630Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4894036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:54:49.4894509Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4894996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:54:49.4895458Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4895647Z 2025-09-07T08:54:49.4895744Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4896208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4896614Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4896931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4897281Z x = module(x) 2025-09-07T08:54:49.4897566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4897897Z out = self.relu(out) 2025-09-07T08:54:49.4897992Z 2025-09-07T08:54:49.4898139Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4898567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4898970Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4899371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:54:49.4899831Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4900295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:54:49.4900759Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4900943Z 2025-09-07T08:54:49.4901037Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4901468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4901878Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4902182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4902491Z x = module(x) 2025-09-07T08:54:49.4902780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4903093Z out = self.relu(out) 2025-09-07T08:54:49.4903191Z 2025-09-07T08:54:49.4903286Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4903714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4904115Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4904427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4904731Z x = module(x) 2025-09-07T08:54:49.4905008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4905316Z out = self.relu(out) 2025-09-07T08:54:49.4905418Z 2025-09-07T08:54:49.4905514Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4905944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4906339Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4906652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4906964Z x = module(x) 2025-09-07T08:54:49.4907251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4907566Z out = self.relu(out) 2025-09-07T08:54:49.4907661Z 2025-09-07T08:54:49.4907736Z cudagraph partition due to non gpu ops 2025-09-07T08:54:49.4907964Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4908392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4908796Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4909138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4909444Z x = module(x) 2025-09-07T08:54:49.4909737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4910070Z out = self.relu(out) 2025-09-07T08:54:49.4910167Z 2025-09-07T08:54:49.4910272Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4910711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4911116Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4911430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4911735Z x = module(x) 2025-09-07T08:54:49.4912022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:54:49.4912327Z out = self.relu(out) 2025-09-07T08:54:49.4912432Z 2025-09-07T08:54:49.4912530Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4912957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4913360Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4913661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4913968Z x = module(x) 2025-09-07T08:54:49.4914252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4914568Z out = self.relu(out) 2025-09-07T08:54:49.4914665Z 2025-09-07T08:54:49.4914768Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4915191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:54:49.4915593Z feats = self.feat_extractor(x) 2025-09-07T08:54:49.4915904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:54:49.4916213Z x = module(x) 2025-09-07T08:54:49.4916488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:54:49.4916804Z out = self.relu(out) 2025-09-07T08:54:49.4916906Z 2025-09-07T08:54:49.4916980Z cudagraph partition due to non gpu ops 2025-09-07T08:54:49.4917203Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4917639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4918044Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4918450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:54:49.4918915Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4919392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:54:49.4919866Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4920050Z 2025-09-07T08:54:49.4920147Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4920582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4920991Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4921391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:54:49.4921891Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4922350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:54:49.4922832Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:54:49.4923016Z 2025-09-07T08:54:49.4923112Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4923574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4923978Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4924367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 86, in forward 2025-09-07T08:54:49.4924782Z out.append(self.upsample(out[-1]) + t) 2025-09-07T08:54:49.4924923Z 2025-09-07T08:54:49.4925021Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4925454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4925860Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4926254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:54:49.4926752Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4927220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:54:49.4927686Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4927853Z 2025-09-07T08:54:49.4927959Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4928387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4928790Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4929192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 86, in forward 2025-09-07T08:54:49.4929605Z out.append(self.upsample(out[-1]) + t) 2025-09-07T08:54:49.4929733Z 2025-09-07T08:54:49.4929838Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4930263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4930669Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4931068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:54:49.4931521Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4931981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:54:49.4932428Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4932602Z 2025-09-07T08:54:49.4932697Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4933129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4933538Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4933936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:54:49.4934381Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4934861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:54:49.4935320Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4935496Z 2025-09-07T08:54:49.4935594Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4936040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4936455Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4936840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:54:49.4937287Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4937739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:54:49.4938198Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4938359Z 2025-09-07T08:54:49.4938464Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4938886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4939287Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4939678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:54:49.4940125Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4940578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:54:49.4941020Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4941189Z 2025-09-07T08:54:49.4941289Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4941719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4942120Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4942519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:54:49.4942955Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4943409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:54:49.4943856Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4944015Z 2025-09-07T08:54:49.4944124Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4944557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4944952Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4945347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:54:49.4945796Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4946258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:54:49.4946710Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4946871Z 2025-09-07T08:54:49.4946975Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4947402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:54:49.4947838Z feat_concat = self.fpn(feats) 2025-09-07T08:54:49.4948231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:54:49.4948678Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4949150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:54:49.4949614Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:54:49.4949781Z 2025-09-07T08:54:49.4949859Z cudagraph partition due to non gpu ops 2025-09-07T08:54:49.4950064Z cudagraph partition due to non gpu ops 2025-09-07T08:54:49.4950287Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4950711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:54:49.4951128Z logits = self.prob_head(feat_concat) 2025-09-07T08:54:49.4951260Z 2025-09-07T08:54:49.4951356Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4951789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:54:49.4952204Z logits = self.prob_head(feat_concat) 2025-09-07T08:54:49.4952328Z 2025-09-07T08:54:49.4952426Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4952851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:54:49.4953256Z logits = self.prob_head(feat_concat) 2025-09-07T08:54:49.4953376Z 2025-09-07T08:54:49.4953478Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4953907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:54:49.4954306Z logits = self.prob_head(feat_concat) 2025-09-07T08:54:49.4954435Z 2025-09-07T08:54:49.4954530Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4954956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:54:49.4955360Z logits = self.prob_head(feat_concat) 2025-09-07T08:54:49.4955482Z 2025-09-07T08:54:49.4955583Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:54:49.4956001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:54:49.4956407Z logits = self.prob_head(feat_concat) 2025-09-07T08:54:49.4956536Z 2025-09-07T08:54:49.4956610Z cudagraph partition due to non gpu ops 2025-09-07T08:55:04.7928230Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:55:04.7929039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/core.py", line 95, in __call__ 2025-09-07T08:55:04.7929699Z for bmap in (proba_map >= self.bin_thresh).astype(np.uint8) 2025-09-07T08:55:04.7929965Z 2025-09-07T08:55:05.6850765Z 2025-09-07T08:55:05.8069906Z running benchmark: 0% 0/30 [00:00blqk", from_blocked_mask[:, 2:-2], exp_blocked_to_pad) 2025-09-07T08:59:04.6113493Z 2025-09-07T08:59:06.2555680Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:06.2556291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:59:06.2556759Z self.query(hidden_states) 2025-09-07T08:59:06.2556911Z 2025-09-07T08:59:06.2557008Z cudagraph partition due to non gpu ops 2025-09-07T08:59:06.2557297Z cudagraph partition due to non gpu ops 2025-09-07T08:59:06.2557515Z cudagraph partition due to non gpu ops 2025-09-07T08:59:06.2557710Z cudagraph partition due to non gpu ops 2025-09-07T08:59:06.2557906Z cudagraph partition due to non gpu ops 2025-09-07T08:59:06.2558101Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9500429Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9501067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 602, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9501666Z first_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, 0], key_layer, ndim=4) 2025-09-07T08:59:08.9502188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:59:08.9502589Z return torch.bmm( 2025-09-07T08:59:08.9502688Z 2025-09-07T08:59:08.9502787Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9502992Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9503188Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9503382Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9503575Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9503764Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9503990Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9504190Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9504375Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9504572Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9504760Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9504951Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9505136Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9505330Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9505522Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9505715Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9506118Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9506649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 647, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9507213Z to_mask.new_ones([bsz, 1, 1, n_rand_blocks * to_block_size]), 2025-09-07T08:59:08.9507382Z 2025-09-07T08:59:08.9507482Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9508018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 653, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9508544Z rand_mask.new_ones([bsz, n_heads, from_block_size, 4 * to_block_size]), 2025-09-07T08:59:08.9508721Z 2025-09-07T08:59:08.9508795Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9508991Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9509254Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9509748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 698, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9510214Z first_band_product = torch.einsum( 2025-09-07T08:59:08.9510337Z 2025-09-07T08:59:08.9510408Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9510602Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9510795Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9511008Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9511489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 687, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9512066Z inner_band_product = self.torch_bmm_nd_transpose(middle_query_matrix, exp_blocked_key_matrix, ndim=5) 2025-09-07T08:59:08.9512592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 494, in torch_bmm_nd_transpose 2025-09-07T08:59:08.9513086Z inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:]).transpose(1, 2) 2025-09-07T08:59:08.9513290Z 2025-09-07T08:59:08.9513397Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9513931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 704, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9514401Z last_band_product = torch.einsum( 2025-09-07T08:59:08.9514523Z 2025-09-07T08:59:08.9514602Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9514791Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9514987Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9515182Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9515378Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9515566Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9515758Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9515951Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9516148Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9516335Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9516531Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9516753Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9517245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 807, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9517802Z last_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, -1], key_layer, ndim=4) 2025-09-07T08:59:08.9518299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:59:08.9518712Z return torch.bmm( 2025-09-07T08:59:08.9518816Z 2025-09-07T08:59:08.9518913Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9519400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 611, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9519962Z first_context_layer = self.torch_bmm_nd(first_attn_weights, value_layer, ndim=4) 2025-09-07T08:59:08.9520437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:59:08.9520922Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:59:08.9521149Z 2025-09-07T08:59:08.9521246Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9521755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 574, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9522237Z rand_attn = np.stack(rand_attn, axis=0) 2025-09-07T08:59:08.9522370Z 2025-09-07T08:59:08.9522464Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9522962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 575, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9523508Z rand_attn = torch.tensor(rand_attn, device=query_layer.device, dtype=torch.long) 2025-09-07T08:59:08.9523711Z 2025-09-07T08:59:08.9523804Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9524300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 620, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9524766Z second_key_mat = torch.cat( 2025-09-07T08:59:08.9524880Z 2025-09-07T08:59:08.9524978Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9525472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 642, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9526053Z second_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, 1], second_key_mat, ndim=4) 2025-09-07T08:59:08.9526572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:59:08.9526967Z return torch.bmm( 2025-09-07T08:59:08.9527057Z 2025-09-07T08:59:08.9527151Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9527639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 643, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9528104Z second_seq_pad = torch.cat( 2025-09-07T08:59:08.9528216Z 2025-09-07T08:59:08.9528318Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9528805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 651, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9529267Z second_rand_pad = torch.cat( 2025-09-07T08:59:08.9529387Z 2025-09-07T08:59:08.9529482Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9529971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 630, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9530433Z second_value_mat = torch.cat( 2025-09-07T08:59:08.9530551Z 2025-09-07T08:59:08.9530653Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9531156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 665, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9531711Z second_context_layer = self.torch_bmm_nd(second_attn_weights, second_value_mat, ndim=4) 2025-09-07T08:59:08.9532180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:59:08.9532673Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:59:08.9532905Z 2025-09-07T08:59:08.9533011Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9533496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 676, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9533960Z exp_blocked_key_matrix = torch.cat( 2025-09-07T08:59:08.9534093Z 2025-09-07T08:59:08.9534221Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9534709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 687, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9535284Z inner_band_product = self.torch_bmm_nd_transpose(middle_query_matrix, exp_blocked_key_matrix, ndim=5) 2025-09-07T08:59:08.9535802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:59:08.9536194Z return torch.bmm( 2025-09-07T08:59:08.9536294Z 2025-09-07T08:59:08.9536387Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9536875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 693, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9537450Z rand_band_product = self.torch_bmm_nd_transpose(middle_query_matrix, gathered_key[:, :, 1:-1], ndim=5) 2025-09-07T08:59:08.9537963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:59:08.9538343Z return torch.bmm( 2025-09-07T08:59:08.9538442Z 2025-09-07T08:59:08.9538537Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9539028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 711, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9539584Z first_band_product += (1.0 - to_mask[:, :, :, :to_block_size].unsqueeze(3)) * attn_mask_penalty 2025-09-07T08:59:08.9539785Z 2025-09-07T08:59:08.9539891Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9540380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 710, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9540876Z inner_band_product += (1.0 - band_mask) * attn_mask_penalty 2025-09-07T08:59:08.9541044Z 2025-09-07T08:59:08.9541141Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9541625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 713, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9542143Z rand_band_product += (1.0 - rand_mask[:, :, 1:-1]) * attn_mask_penalty 2025-09-07T08:59:08.9542314Z 2025-09-07T08:59:08.9542420Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9542902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 712, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9543443Z last_band_product += (1.0 - to_mask[:, :, :, -to_block_size:].unsqueeze(3)) * attn_mask_penalty 2025-09-07T08:59:08.9543664Z 2025-09-07T08:59:08.9543762Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9544253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 716, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9544729Z band_product = torch.cat( 2025-09-07T08:59:08.9544839Z 2025-09-07T08:59:08.9544941Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9545453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 679, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9545926Z exp_blocked_value_matrix = torch.cat( 2025-09-07T08:59:08.9546062Z 2025-09-07T08:59:08.9546156Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9546656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 727, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9547124Z context_layer = self.torch_bmm_nd( 2025-09-07T08:59:08.9547503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:59:08.9547995Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:59:08.9548224Z 2025-09-07T08:59:08.9548318Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9548808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 734, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9549276Z context_layer += self.torch_bmm_nd( 2025-09-07T08:59:08.9549660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:59:08.9550146Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:59:08.9550366Z 2025-09-07T08:59:08.9550462Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9550953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 740, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9551423Z context_layer += torch.einsum( 2025-09-07T08:59:08.9551543Z 2025-09-07T08:59:08.9551639Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9552127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 743, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9552595Z context_layer += torch.einsum( 2025-09-07T08:59:08.9552715Z 2025-09-07T08:59:08.9552820Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9553311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 754, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9553773Z second_last_key_mat = torch.cat( 2025-09-07T08:59:08.9553904Z 2025-09-07T08:59:08.9554002Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9554496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 776, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9555088Z second_last_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, -2], second_last_key_mat, ndim=4) 2025-09-07T08:59:08.9555623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:59:08.9556021Z return torch.bmm( 2025-09-07T08:59:08.9556114Z 2025-09-07T08:59:08.9556226Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9556717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 777, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9557200Z second_last_seq_pad = torch.cat( 2025-09-07T08:59:08.9557318Z 2025-09-07T08:59:08.9557421Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9557919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 785, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9558377Z second_last_rand_pad = torch.cat( 2025-09-07T08:59:08.9558507Z 2025-09-07T08:59:08.9558601Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9559102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 764, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9559573Z second_last_value_mat = torch.cat( 2025-09-07T08:59:08.9559698Z 2025-09-07T08:59:08.9559809Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9560287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 799, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9560871Z second_last_context_layer = self.torch_bmm_nd(second_last_attn_weights, second_last_value_mat, ndim=4) 2025-09-07T08:59:08.9561365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:59:08.9561848Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:59:08.9562063Z 2025-09-07T08:59:08.9562172Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9562667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 813, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9563198Z last_context_layer = self.torch_bmm_nd(last_attn_weights, value_layer, ndim=4) 2025-09-07T08:59:08.9563658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:59:08.9564138Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:59:08.9564353Z 2025-09-07T08:59:08.9564441Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9564652Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9564874Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9565370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 743, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9565838Z context_layer += torch.einsum( 2025-09-07T08:59:08.9565958Z 2025-09-07T08:59:08.9566041Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9566237Z cudagraph partition due to non gpu ops 2025-09-07T08:59:08.9566463Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9566953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 817, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9567421Z context_layer = torch.cat( 2025-09-07T08:59:08.9567532Z 2025-09-07T08:59:08.9567639Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:08.9568128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 821, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:59:08.9568920Z context_layer = context_layer.view((bsz, n_heads, from_seq_len, -1)) * from_mask 2025-09-07T08:59:08.9569133Z 2025-09-07T08:59:16.4478217Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:16.4478965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 478, in torch_dynamo_resume_in_forward_at_455 2025-09-07T08:59:16.4479799Z context_layer = context_layer.contiguous().view(batch_size, from_seq_length, -1) 2025-09-07T08:59:16.4480085Z 2025-09-07T08:59:17.1326068Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:17.1326752Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:59:17.1327272Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:59:17.1327952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:59:17.1328394Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:17.1328538Z 2025-09-07T08:59:17.1328623Z cudagraph partition due to non gpu ops 2025-09-07T08:59:17.1328840Z cudagraph partition due to non gpu ops 2025-09-07T08:59:18.0679753Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:18.0680386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:18.0680870Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:18.0681275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:18.0681661Z return forward_fn(*input_tensors) 2025-09-07T08:59:18.0682059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:18.0682529Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:18.0682956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:59:18.0683342Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:18.0683480Z 2025-09-07T08:59:18.0683571Z cudagraph partition due to non gpu ops 2025-09-07T08:59:18.0683770Z cudagraph partition due to non gpu ops 2025-09-07T08:59:18.0684000Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:18.0684464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:18.0684904Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:18.0685275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:18.0685653Z return forward_fn(*input_tensors) 2025-09-07T08:59:18.0686045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:18.0686488Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:18.0686902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:59:18.0687307Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:59:18.0687675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:59:18.0688103Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:59:18.0688321Z 2025-09-07T08:59:18.0688405Z cudagraph partition due to non gpu ops 2025-09-07T08:59:18.0688608Z cudagraph partition due to non gpu ops 2025-09-07T08:59:18.9504321Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:18.9504924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:59:18.9505320Z self.query(hidden_states) 2025-09-07T08:59:18.9505441Z 2025-09-07T08:59:18.9505597Z cudagraph partition due to non gpu ops 2025-09-07T08:59:18.9505804Z cudagraph partition due to non gpu ops 2025-09-07T08:59:18.9506065Z cudagraph partition due to non gpu ops 2025-09-07T08:59:18.9506270Z cudagraph partition due to non gpu ops 2025-09-07T08:59:18.9506469Z cudagraph partition due to non gpu ops 2025-09-07T08:59:18.9506660Z cudagraph partition due to non gpu ops 2025-09-07T08:59:19.8360117Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:19.8360717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:59:19.8361488Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:59:19.8361956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:59:19.8362361Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:19.8362518Z 2025-09-07T08:59:19.8362602Z cudagraph partition due to non gpu ops 2025-09-07T08:59:19.8362813Z cudagraph partition due to non gpu ops 2025-09-07T08:59:20.0699832Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:20.0700378Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:20.0700848Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:20.0701246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:20.0701652Z return forward_fn(*input_tensors) 2025-09-07T08:59:20.0702060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:20.0702522Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:20.0702944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:59:20.0703340Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:20.0703484Z 2025-09-07T08:59:20.0703568Z cudagraph partition due to non gpu ops 2025-09-07T08:59:20.0703778Z cudagraph partition due to non gpu ops 2025-09-07T08:59:20.0704007Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:20.0704459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:20.0704901Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:20.0705283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:20.0705654Z return forward_fn(*input_tensors) 2025-09-07T08:59:20.0706045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:20.0706485Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:20.0706909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:59:20.0707318Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:59:20.0707687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:59:20.0708116Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:59:20.0708505Z 2025-09-07T08:59:20.0708590Z cudagraph partition due to non gpu ops 2025-09-07T08:59:20.0708797Z cudagraph partition due to non gpu ops 2025-09-07T08:59:20.6324795Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:20.6325391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:59:20.6326084Z self.query(hidden_states) 2025-09-07T08:59:20.6326268Z 2025-09-07T08:59:20.6326351Z cudagraph partition due to non gpu ops 2025-09-07T08:59:20.6326559Z cudagraph partition due to non gpu ops 2025-09-07T08:59:20.6326754Z cudagraph partition due to non gpu ops 2025-09-07T08:59:20.6326949Z cudagraph partition due to non gpu ops 2025-09-07T08:59:20.6327137Z cudagraph partition due to non gpu ops 2025-09-07T08:59:20.6327327Z cudagraph partition due to non gpu ops 2025-09-07T08:59:21.5250673Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:21.5251630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:59:21.5252171Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:59:21.5252611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:59:21.5253025Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:21.5253175Z 2025-09-07T08:59:21.5253261Z cudagraph partition due to non gpu ops 2025-09-07T08:59:21.5253469Z cudagraph partition due to non gpu ops 2025-09-07T08:59:21.7596494Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:21.7597165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:21.7597745Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:21.7598240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:21.7598630Z return forward_fn(*input_tensors) 2025-09-07T08:59:21.7599039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:21.7599496Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:21.7599923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:59:21.7600309Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:21.7600442Z 2025-09-07T08:59:21.7600521Z cudagraph partition due to non gpu ops 2025-09-07T08:59:21.7600729Z cudagraph partition due to non gpu ops 2025-09-07T08:59:21.7600955Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:21.7601418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:21.7601864Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:21.7602249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:21.7602624Z return forward_fn(*input_tensors) 2025-09-07T08:59:21.7603021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:21.7603462Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:21.7603874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:59:21.7604280Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:59:21.7604820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:59:21.7605267Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:59:21.7605491Z 2025-09-07T08:59:21.7605576Z cudagraph partition due to non gpu ops 2025-09-07T08:59:21.7605842Z cudagraph partition due to non gpu ops 2025-09-07T08:59:22.0280703Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:22.0281516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:59:22.0281997Z self.query(hidden_states) 2025-09-07T08:59:22.0282134Z 2025-09-07T08:59:22.0282242Z cudagraph partition due to non gpu ops 2025-09-07T08:59:22.0282498Z cudagraph partition due to non gpu ops 2025-09-07T08:59:22.0282748Z cudagraph partition due to non gpu ops 2025-09-07T08:59:22.0282979Z cudagraph partition due to non gpu ops 2025-09-07T08:59:22.0283232Z cudagraph partition due to non gpu ops 2025-09-07T08:59:22.0283433Z cudagraph partition due to non gpu ops 2025-09-07T08:59:22.9064523Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:22.9065200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:59:22.9065846Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:59:22.9066322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:59:22.9066717Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:22.9066854Z 2025-09-07T08:59:22.9066943Z cudagraph partition due to non gpu ops 2025-09-07T08:59:22.9067146Z cudagraph partition due to non gpu ops 2025-09-07T08:59:23.1425531Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:23.1427247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:23.1427855Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:23.1428412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:23.1428816Z return forward_fn(*input_tensors) 2025-09-07T08:59:23.1429220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:23.1429679Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:23.1430095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:59:23.1430483Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:23.1430626Z 2025-09-07T08:59:23.1430712Z cudagraph partition due to non gpu ops 2025-09-07T08:59:23.1430924Z cudagraph partition due to non gpu ops 2025-09-07T08:59:23.1431149Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:23.1431629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:23.1432079Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:23.1432466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:23.1432849Z return forward_fn(*input_tensors) 2025-09-07T08:59:23.1433248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:23.1433685Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:23.1434278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:59:23.1434697Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:59:23.1435069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:59:23.1435559Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:59:23.1435781Z 2025-09-07T08:59:23.1435895Z cudagraph partition due to non gpu ops 2025-09-07T08:59:23.1436098Z cudagraph partition due to non gpu ops 2025-09-07T08:59:23.4142738Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:23.4143300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:59:23.4143745Z self.query(hidden_states) 2025-09-07T08:59:23.4143888Z 2025-09-07T08:59:23.4143983Z cudagraph partition due to non gpu ops 2025-09-07T08:59:23.4144421Z cudagraph partition due to non gpu ops 2025-09-07T08:59:23.4144670Z cudagraph partition due to non gpu ops 2025-09-07T08:59:23.4144901Z cudagraph partition due to non gpu ops 2025-09-07T08:59:23.4145126Z cudagraph partition due to non gpu ops 2025-09-07T08:59:23.4145358Z cudagraph partition due to non gpu ops 2025-09-07T08:59:24.3029620Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:24.3030319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:59:24.3030952Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:59:24.3031475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:59:24.3031894Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:24.3032040Z 2025-09-07T08:59:24.3032130Z cudagraph partition due to non gpu ops 2025-09-07T08:59:24.3032347Z cudagraph partition due to non gpu ops 2025-09-07T08:59:24.5394259Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:24.5394837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:24.5395335Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:24.5395725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:24.5396128Z return forward_fn(*input_tensors) 2025-09-07T08:59:24.5396533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:24.5396978Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:24.5397410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:59:24.5397795Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:24.5397939Z 2025-09-07T08:59:24.5398019Z cudagraph partition due to non gpu ops 2025-09-07T08:59:24.5398243Z cudagraph partition due to non gpu ops 2025-09-07T08:59:24.5398475Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:24.5398938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:24.5399385Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:24.5399769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:24.5400138Z return forward_fn(*input_tensors) 2025-09-07T08:59:24.5400530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:24.5401177Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:24.5401602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:59:24.5402012Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:59:24.5402439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:59:24.5402918Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:59:24.5403140Z 2025-09-07T08:59:24.5403224Z cudagraph partition due to non gpu ops 2025-09-07T08:59:24.5403419Z cudagraph partition due to non gpu ops 2025-09-07T08:59:24.8091634Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:24.8092437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:59:24.8092976Z self.query(hidden_states) 2025-09-07T08:59:24.8093097Z 2025-09-07T08:59:24.8093194Z cudagraph partition due to non gpu ops 2025-09-07T08:59:24.8093407Z cudagraph partition due to non gpu ops 2025-09-07T08:59:24.8093600Z cudagraph partition due to non gpu ops 2025-09-07T08:59:24.8093802Z cudagraph partition due to non gpu ops 2025-09-07T08:59:24.8093998Z cudagraph partition due to non gpu ops 2025-09-07T08:59:24.8094197Z cudagraph partition due to non gpu ops 2025-09-07T08:59:25.6921453Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:25.6922081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:59:25.6922598Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:59:25.6923063Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:59:25.6923471Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:25.6923620Z 2025-09-07T08:59:25.6923703Z cudagraph partition due to non gpu ops 2025-09-07T08:59:25.6923912Z cudagraph partition due to non gpu ops 2025-09-07T08:59:25.9271794Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:25.9272557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:25.9273148Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:25.9273641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:25.9274099Z return forward_fn(*input_tensors) 2025-09-07T08:59:25.9274525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:25.9274975Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:25.9275389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:59:25.9275776Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:25.9275920Z 2025-09-07T08:59:25.9276002Z cudagraph partition due to non gpu ops 2025-09-07T08:59:25.9276209Z cudagraph partition due to non gpu ops 2025-09-07T08:59:25.9276436Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:25.9276896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:25.9277343Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:25.9277726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:25.9278307Z return forward_fn(*input_tensors) 2025-09-07T08:59:25.9278715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:25.9279159Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:25.9281208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:59:25.9281669Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:59:25.9282038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:59:25.9282459Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:59:25.9282685Z 2025-09-07T08:59:25.9282762Z cudagraph partition due to non gpu ops 2025-09-07T08:59:25.9282964Z cudagraph partition due to non gpu ops 2025-09-07T08:59:26.1986133Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:26.1986823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:59:26.1987317Z self.query(hidden_states) 2025-09-07T08:59:26.1987467Z 2025-09-07T08:59:26.1987565Z cudagraph partition due to non gpu ops 2025-09-07T08:59:26.1987818Z cudagraph partition due to non gpu ops 2025-09-07T08:59:26.1988069Z cudagraph partition due to non gpu ops 2025-09-07T08:59:26.1988301Z cudagraph partition due to non gpu ops 2025-09-07T08:59:26.1988500Z cudagraph partition due to non gpu ops 2025-09-07T08:59:26.1988694Z cudagraph partition due to non gpu ops 2025-09-07T08:59:27.0807759Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:27.0808332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:59:27.0808872Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:59:27.0809306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:59:27.0809699Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:27.0809856Z 2025-09-07T08:59:27.0809941Z cudagraph partition due to non gpu ops 2025-09-07T08:59:27.0810157Z cudagraph partition due to non gpu ops 2025-09-07T08:59:27.3153484Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:27.3154037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:27.3154505Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:27.3154909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:27.3155312Z return forward_fn(*input_tensors) 2025-09-07T08:59:27.3155716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:27.3156161Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:27.3156587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:59:27.3156981Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:27.3157118Z 2025-09-07T08:59:27.3157202Z cudagraph partition due to non gpu ops 2025-09-07T08:59:27.3157409Z cudagraph partition due to non gpu ops 2025-09-07T08:59:27.3157637Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:27.3158099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:27.3158757Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:27.3159149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:27.3159522Z return forward_fn(*input_tensors) 2025-09-07T08:59:27.3159918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:27.3160435Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:27.3160891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:59:27.3161309Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:59:27.3161679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:59:27.3162164Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:59:27.3162385Z 2025-09-07T08:59:27.3162471Z cudagraph partition due to non gpu ops 2025-09-07T08:59:27.3162665Z cudagraph partition due to non gpu ops 2025-09-07T08:59:27.5842508Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:27.5843185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:59:27.5843622Z self.query(hidden_states) 2025-09-07T08:59:27.5843753Z 2025-09-07T08:59:27.5843839Z cudagraph partition due to non gpu ops 2025-09-07T08:59:27.5844050Z cudagraph partition due to non gpu ops 2025-09-07T08:59:27.5844240Z cudagraph partition due to non gpu ops 2025-09-07T08:59:27.5844440Z cudagraph partition due to non gpu ops 2025-09-07T08:59:27.5844635Z cudagraph partition due to non gpu ops 2025-09-07T08:59:27.5844831Z cudagraph partition due to non gpu ops 2025-09-07T08:59:28.4643858Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:28.4644543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:59:28.4645050Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:59:28.4645496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:59:28.4645891Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:28.4646039Z 2025-09-07T08:59:28.4646124Z cudagraph partition due to non gpu ops 2025-09-07T08:59:28.4646336Z cudagraph partition due to non gpu ops 2025-09-07T08:59:28.7034690Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:28.7035358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:28.7035939Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:28.7036367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:28.7036755Z return forward_fn(*input_tensors) 2025-09-07T08:59:28.7037166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:28.7037626Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:28.7038058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:59:28.7038438Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:28.7038581Z 2025-09-07T08:59:28.7038661Z cudagraph partition due to non gpu ops 2025-09-07T08:59:28.7038868Z cudagraph partition due to non gpu ops 2025-09-07T08:59:28.7039097Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:28.7039740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:59:28.7040196Z layer_output = apply_chunking_to_forward( 2025-09-07T08:59:28.7040585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:59:28.7041016Z return forward_fn(*input_tensors) 2025-09-07T08:59:28.7041457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:59:28.7041899Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:59:28.7042310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:59:28.7042722Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:59:28.7043144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:59:28.7043582Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:59:28.7043801Z 2025-09-07T08:59:28.7043885Z cudagraph partition due to non gpu ops 2025-09-07T08:59:28.7044081Z cudagraph partition due to non gpu ops 2025-09-07T08:59:28.7776874Z W0907 08:59:28.777225 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] torch._dynamo hit config.recompile_limit (8) 2025-09-07T08:59:28.7777801Z W0907 08:59:28.777225 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] function: 'forward' (/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py:417) 2025-09-07T08:59:28.7779556Z W0907 08:59:28.777225 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] last reason: 10/7: self.seed == 7 # context_layer, attention_probs = self.bigbird_block_sparse_attention( # transformers/models/big_bird/modeling_big_bird.py:455 in forward (HINT: torch.compile considers integer attributes of the nn.Module to be static. If you are observing recompilation, you might want to make this integer dynamic using torch._dynamo.config.allow_unspec_int_on_nn_module = True, or convert this integer into a tensor.) 2025-09-07T08:59:28.7781127Z W0907 08:59:28.777225 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] To log all recompilation reasons, use TORCH_LOGS="recompiles". 2025-09-07T08:59:28.7781950Z W0907 08:59:28.777225 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html 2025-09-07T08:59:28.7996365Z W0907 08:59:28.798521 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] torch._dynamo hit config.recompile_limit (8) 2025-09-07T08:59:28.7997473Z W0907 08:59:28.798521 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] function: 'torch_dynamo_resume_in_forward_at_1381' (/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py:1381) 2025-09-07T08:59:28.7998525Z W0907 08:59:28.798521 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] last reason: 17/7: ___check_obj_id(self._modules['output']._modules['dense']._parameters['bias'], 139628094516976) # return F.linear(input, self.weight, self.bias) # nn/modules/linear.py:134 in forward 2025-09-07T08:59:28.7999357Z W0907 08:59:28.798521 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] To log all recompilation reasons, use TORCH_LOGS="recompiles". 2025-09-07T08:59:28.8000037Z W0907 08:59:28.798521 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [17/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html 2025-09-07T08:59:28.8022226Z W0907 08:59:28.801833 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] torch._dynamo hit config.recompile_limit (8) 2025-09-07T08:59:28.8023115Z W0907 08:59:28.801833 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] function: 'torch_dynamo_resume_in_forward_at_1468' (/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py:1468) 2025-09-07T08:59:28.8024349Z W0907 08:59:28.801833 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] last reason: 18/7: ___check_obj_id(self._modules['output']._modules['dense']._parameters['bias'], 139628300879408) # return F.linear(input, self.weight, self.bias) # nn/modules/linear.py:134 in forward 2025-09-07T08:59:28.8025343Z W0907 08:59:28.801833 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] To log all recompilation reasons, use TORCH_LOGS="recompiles". 2025-09-07T08:59:28.8026186Z W0907 08:59:28.801833 159670 site-packages/torch/_dynamo/convert_frame.py:1358] [18/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html 2025-09-07T08:59:29.0081184Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:29.0081867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2010, in torch_dynamo_resume_in_forward_at_1991 2025-09-07T08:59:29.0082579Z pooler_output = self.activation(self.pooler(sequence_output[:, 0, :])) if (self.pooler is not None) else None 2025-09-07T08:59:29.0082872Z 2025-09-07T08:59:29.0082987Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:29.0083533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2010, in torch_dynamo_resume_in_forward_at_1991 2025-09-07T08:59:29.0084185Z pooler_output = self.activation(self.pooler(sequence_output[:, 0, :])) if (self.pooler is not None) else None 2025-09-07T08:59:29.0084480Z 2025-09-07T08:59:29.0084600Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:29.0085069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2010, in torch_dynamo_resume_in_forward_at_1991 2025-09-07T08:59:29.0085631Z pooler_output = self.activation(self.pooler(sequence_output[:, 0, :])) if (self.pooler is not None) else None 2025-09-07T08:59:29.0085893Z 2025-09-07T08:59:29.7640757Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:59:29.7641437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2318, in torch_dynamo_resume_in_forward_at_2303 2025-09-07T08:59:29.7642016Z prediction_scores = self.cls(sequence_output) 2025-09-07T08:59:29.7642455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1682, in forward 2025-09-07T08:59:29.7642884Z prediction_scores = self.predictions(sequence_output) 2025-09-07T08:59:29.7643317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1670, in forward 2025-09-07T08:59:29.7643718Z hidden_states = self.transform(hidden_states) 2025-09-07T08:59:29.7644106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1645, in forward 2025-09-07T08:59:29.7644505Z hidden_states = self.dense(hidden_states) 2025-09-07T08:59:29.7644646Z 2025-09-07T08:59:29.7644735Z cudagraph partition due to non gpu ops 2025-09-07T08:59:29.7644945Z cudagraph partition due to non gpu ops 2025-09-07T08:59:29.7645149Z cudagraph partition due to non gpu ops 2025-09-07T08:59:29.7645337Z cudagraph partition due to non gpu ops 2025-09-07T08:59:31.8769058Z 2025-09-07T08:59:32.3248581Z running benchmark: 0% 0/30 [00:00bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0202054Z 2025-09-07T09:02:14.0202133Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0202342Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0202572Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0203059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0203528Z layer_outputs = layer_module( 2025-09-07T09:02:14.0203872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0204219Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0204618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0205004Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0205404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0205802Z self_outputs = self.self( 2025-09-07T09:02:14.0206182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0206608Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0207078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0207636Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0207873Z 2025-09-07T09:02:14.0207949Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0208149Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0208355Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0208547Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0208771Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0209293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0209763Z layer_outputs = layer_module( 2025-09-07T09:02:14.0210099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0210475Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0210878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0211289Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0211687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0212083Z self_outputs = self.self( 2025-09-07T09:02:14.0212490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T09:02:14.0212931Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0213425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0213964Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T09:02:14.0214459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T09:02:14.0214972Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T09:02:14.0215175Z 2025-09-07T09:02:14.0215253Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0215487Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0215992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0216465Z layer_outputs = layer_module( 2025-09-07T09:02:14.0216803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0217158Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0217560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0217953Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0218344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0218737Z self_outputs = self.self( 2025-09-07T09:02:14.0219122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:02:14.0219515Z attn_scores += diagonal_mask 2025-09-07T09:02:14.0219635Z 2025-09-07T09:02:14.0219737Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0220243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0220713Z layer_outputs = layer_module( 2025-09-07T09:02:14.0221047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0221392Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0221783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0222179Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0222578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0222987Z self_outputs = self.self( 2025-09-07T09:02:14.0223364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:02:14.0223763Z attn_probs = nn.functional.softmax( 2025-09-07T09:02:14.0223915Z 2025-09-07T09:02:14.0223990Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0224196Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0224420Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0224913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0225368Z layer_outputs = layer_module( 2025-09-07T09:02:14.0225710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0226098Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0226499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0226896Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0227284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0227693Z self_outputs = self.self( 2025-09-07T09:02:14.0228078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0228518Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0229014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0229579Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:02:14.0229994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0230336Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0230484Z 2025-09-07T09:02:14.0230597Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0231078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0231538Z layer_outputs = layer_module( 2025-09-07T09:02:14.0231874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0232227Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0232632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0233030Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0233432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0233826Z self_outputs = self.self( 2025-09-07T09:02:14.0234211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0234650Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0235142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0235662Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:02:14.0236152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:02:14.0236619Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:02:14.0236950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0237285Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0237440Z 2025-09-07T09:02:14.0237543Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0238056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0238523Z layer_outputs = layer_module( 2025-09-07T09:02:14.0238861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0239205Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0239615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0240033Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0240438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0240836Z self_outputs = self.self( 2025-09-07T09:02:14.0241261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0241697Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0242203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0242736Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0242932Z 2025-09-07T09:02:14.0243039Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0243528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0243995Z layer_outputs = layer_module( 2025-09-07T09:02:14.0244331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0244682Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0245073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0245474Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0245871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0246265Z self_outputs = self.self( 2025-09-07T09:02:14.0246649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0247080Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0247584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0248122Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0248314Z 2025-09-07T09:02:14.0248422Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0248914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0249364Z layer_outputs = layer_module( 2025-09-07T09:02:14.0249698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0250045Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0250465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0250873Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0251547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0251974Z self_outputs = self.self( 2025-09-07T09:02:14.0252362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:02:14.0252868Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:02:14.0253095Z 2025-09-07T09:02:14.0253185Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0253387Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0253591Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0253807Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0254012Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0254204Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0254432Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0254921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0255406Z layer_outputs = layer_module( 2025-09-07T09:02:14.0255742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0256082Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0256487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0256882Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0257284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0257681Z self_outputs = self.self( 2025-09-07T09:02:14.0258054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:02:14.0258462Z query_vectors = self.query(hidden_states) 2025-09-07T09:02:14.0258602Z 2025-09-07T09:02:14.0258676Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0258880Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0259099Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0259592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0260052Z layer_outputs = layer_module( 2025-09-07T09:02:14.0260388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0260738Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0261135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0261531Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0261934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0262325Z self_outputs = self.self( 2025-09-07T09:02:14.0262706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0263122Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0263599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0264204Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0264438Z 2025-09-07T09:02:14.0264524Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0264731Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0264951Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0265461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0265926Z layer_outputs = layer_module( 2025-09-07T09:02:14.0266257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0266597Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0266995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0267410Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0267814Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0268207Z self_outputs = self.self( 2025-09-07T09:02:14.0268784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0270859Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0271340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0271907Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0272139Z 2025-09-07T09:02:14.0272226Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0272429Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0272666Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0273169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0273624Z layer_outputs = layer_module( 2025-09-07T09:02:14.0273958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0274296Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0274688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0275078Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0275463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0275840Z self_outputs = self.self( 2025-09-07T09:02:14.0276213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:02:14.0276599Z attn_scores += diagonal_mask 2025-09-07T09:02:14.0276714Z 2025-09-07T09:02:14.0276820Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0277298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0277736Z layer_outputs = layer_module( 2025-09-07T09:02:14.0278058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0278396Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0278787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0279167Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0279573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0279959Z self_outputs = self.self( 2025-09-07T09:02:14.0280327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:02:14.0280757Z attn_probs = nn.functional.softmax( 2025-09-07T09:02:14.0280881Z 2025-09-07T09:02:14.0280962Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0281154Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0281374Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0281848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0282294Z layer_outputs = layer_module( 2025-09-07T09:02:14.0282635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0282974Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0283366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0283768Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0284155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0284528Z self_outputs = self.self( 2025-09-07T09:02:14.0284903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0285328Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0285818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0286364Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:02:14.0286756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0287094Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0287245Z 2025-09-07T09:02:14.0287345Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0287822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0288271Z layer_outputs = layer_module( 2025-09-07T09:02:14.0288590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0288927Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0289321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0289710Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0290093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0290481Z self_outputs = self.self( 2025-09-07T09:02:14.0290852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0291277Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0291765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0292257Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:02:14.0292744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:02:14.0293191Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:02:14.0293514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0293863Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0294005Z 2025-09-07T09:02:14.0294102Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0294581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0295038Z layer_outputs = layer_module( 2025-09-07T09:02:14.0295370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0295714Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0296120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0296516Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0296905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0297306Z self_outputs = self.self( 2025-09-07T09:02:14.0297678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0298094Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0298578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0299095Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0299286Z 2025-09-07T09:02:14.0299393Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0299867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0300312Z layer_outputs = layer_module( 2025-09-07T09:02:14.0300641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0300983Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0301374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0301762Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0302141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0302527Z self_outputs = self.self( 2025-09-07T09:02:14.0302902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0303327Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0303811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0304328Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0304527Z 2025-09-07T09:02:14.0304624Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0305103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0305558Z layer_outputs = layer_module( 2025-09-07T09:02:14.0305899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0306239Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0306629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0307041Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0307431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0307808Z self_outputs = self.self( 2025-09-07T09:02:14.0308179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:02:14.0308671Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:02:14.0308901Z 2025-09-07T09:02:14.0308977Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0309193Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0309389Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0309584Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0309778Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0309972Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0310206Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0310692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0311146Z layer_outputs = layer_module( 2025-09-07T09:02:14.0311473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0311815Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0312202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0312595Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0312987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0313370Z self_outputs = self.self( 2025-09-07T09:02:14.0313743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:02:14.0314130Z query_vectors = self.query(hidden_states) 2025-09-07T09:02:14.0314267Z 2025-09-07T09:02:14.0314342Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0314534Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0314755Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0315226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0315679Z layer_outputs = layer_module( 2025-09-07T09:02:14.0316000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0316338Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0316724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0317108Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0317499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0317883Z self_outputs = self.self( 2025-09-07T09:02:14.0318252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0318669Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0319146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0319698Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0319931Z 2025-09-07T09:02:14.0320009Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0320220Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0320444Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0320917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0321365Z layer_outputs = layer_module( 2025-09-07T09:02:14.0321690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0322039Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0322440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0322828Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0323210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0323615Z self_outputs = self.self( 2025-09-07T09:02:14.0323991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0324402Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0324875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0325424Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0325650Z 2025-09-07T09:02:14.0325734Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0325936Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0326155Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0326636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0327095Z layer_outputs = layer_module( 2025-09-07T09:02:14.0327429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0327774Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0328160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0328553Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0328945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0329330Z self_outputs = self.self( 2025-09-07T09:02:14.0329697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:02:14.0330085Z attn_scores += diagonal_mask 2025-09-07T09:02:14.0330211Z 2025-09-07T09:02:14.0330311Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0330790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0331242Z layer_outputs = layer_module( 2025-09-07T09:02:14.0331560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0331901Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0332352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0332748Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0333135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0333525Z self_outputs = self.self( 2025-09-07T09:02:14.0333893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:02:14.0334283Z attn_probs = nn.functional.softmax( 2025-09-07T09:02:14.0334405Z 2025-09-07T09:02:14.0334488Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0334686Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0334902Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0335405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0335863Z layer_outputs = layer_module( 2025-09-07T09:02:14.0336192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0336525Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0336930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0337319Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0337701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0338085Z self_outputs = self.self( 2025-09-07T09:02:14.0338445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0338873Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0339360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0339900Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:02:14.0340303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0340623Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0340770Z 2025-09-07T09:02:14.0340869Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0341345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0341789Z layer_outputs = layer_module( 2025-09-07T09:02:14.0342116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0342446Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0342833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0343220Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0343604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0343989Z self_outputs = self.self( 2025-09-07T09:02:14.0344353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0344774Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0345257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0345780Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:02:14.0346255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:02:14.0346687Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:02:14.0347024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0347352Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0347496Z 2025-09-07T09:02:14.0347600Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0348071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0348523Z layer_outputs = layer_module( 2025-09-07T09:02:14.0348863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0349208Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0349599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0349998Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0350389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0350774Z self_outputs = self.self( 2025-09-07T09:02:14.0351152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0351579Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0352072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0352605Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0352807Z 2025-09-07T09:02:14.0352904Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0353390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0353851Z layer_outputs = layer_module( 2025-09-07T09:02:14.0354176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0354521Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0354916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0355306Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0355702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0356083Z self_outputs = self.self( 2025-09-07T09:02:14.0356460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0356888Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0357382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0357905Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0358095Z 2025-09-07T09:02:14.0358193Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0358678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0359145Z layer_outputs = layer_module( 2025-09-07T09:02:14.0359474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0359811Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0360212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0360601Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0360988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0361365Z self_outputs = self.self( 2025-09-07T09:02:14.0361727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:02:14.0362229Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:02:14.0362462Z 2025-09-07T09:02:14.0362539Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0362744Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0362943Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0363130Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0363346Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0363538Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0363761Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0364231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0364676Z layer_outputs = layer_module( 2025-09-07T09:02:14.0365001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0365341Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0365734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0366115Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0366501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0366888Z self_outputs = self.self( 2025-09-07T09:02:14.0367259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:02:14.0367652Z query_vectors = self.query(hidden_states) 2025-09-07T09:02:14.0367782Z 2025-09-07T09:02:14.0367856Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0368052Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0368334Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0368819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0369267Z layer_outputs = layer_module( 2025-09-07T09:02:14.0369592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0369933Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0370322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0370712Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0371097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0371484Z self_outputs = self.self( 2025-09-07T09:02:14.0371862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0372308Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0372777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0373315Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0373579Z 2025-09-07T09:02:14.0373655Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0373855Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0374081Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0374563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0375015Z layer_outputs = layer_module( 2025-09-07T09:02:14.0375379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0375727Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0376126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0376545Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0376940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0377330Z self_outputs = self.self( 2025-09-07T09:02:14.0377706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0378121Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0378574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0379130Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0379363Z 2025-09-07T09:02:14.0379439Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0379638Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0379861Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0380336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0380787Z layer_outputs = layer_module( 2025-09-07T09:02:14.0381114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0381458Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0381851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0382233Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0382620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0383002Z self_outputs = self.self( 2025-09-07T09:02:14.0383376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:02:14.0383754Z attn_scores += diagonal_mask 2025-09-07T09:02:14.0383874Z 2025-09-07T09:02:14.0383969Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0384442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0384888Z layer_outputs = layer_module( 2025-09-07T09:02:14.0385214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0385564Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0385957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0386343Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0386752Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0387135Z self_outputs = self.self( 2025-09-07T09:02:14.0387499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:02:14.0387889Z attn_probs = nn.functional.softmax( 2025-09-07T09:02:14.0388025Z 2025-09-07T09:02:14.0388100Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0388299Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0388515Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0389005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0389456Z layer_outputs = layer_module( 2025-09-07T09:02:14.0389784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0390150Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0390537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0390937Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0391325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0391706Z self_outputs = self.self( 2025-09-07T09:02:14.0392083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0392501Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0392993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0393538Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:02:14.0393938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0394271Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0394413Z 2025-09-07T09:02:14.0394511Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0394991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0395448Z layer_outputs = layer_module( 2025-09-07T09:02:14.0395774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0396114Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0396497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0396890Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0397279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0397663Z self_outputs = self.self( 2025-09-07T09:02:14.0398033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0398451Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0398960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0399468Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:02:14.0399941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:02:14.0400408Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:02:14.0400727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0401064Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0401216Z 2025-09-07T09:02:14.0401316Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0401821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0402277Z layer_outputs = layer_module( 2025-09-07T09:02:14.0402600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0402939Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0403347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0403736Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0404112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0404492Z self_outputs = self.self( 2025-09-07T09:02:14.0404861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0405282Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0405771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0406289Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0406490Z 2025-09-07T09:02:14.0406589Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0407065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0407513Z layer_outputs = layer_module( 2025-09-07T09:02:14.0407844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0408174Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0408565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0408954Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0409343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0409719Z self_outputs = self.self( 2025-09-07T09:02:14.0410088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0410505Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0410986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0411507Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0411695Z 2025-09-07T09:02:14.0411802Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0412285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0412739Z layer_outputs = layer_module( 2025-09-07T09:02:14.0413066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0413427Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0413824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0414212Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0414607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0414992Z self_outputs = self.self( 2025-09-07T09:02:14.0415388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:02:14.0415881Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:02:14.0416108Z 2025-09-07T09:02:14.0416184Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0416406Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0416611Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0416814Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0417005Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0417204Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0417433Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0417922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0418373Z layer_outputs = layer_module( 2025-09-07T09:02:14.0418708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0419056Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0419454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0419855Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0420242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0420633Z self_outputs = self.self( 2025-09-07T09:02:14.0421012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:02:14.0421418Z query_vectors = self.query(hidden_states) 2025-09-07T09:02:14.0421552Z 2025-09-07T09:02:14.0421635Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0421832Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0422062Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0422544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0423008Z layer_outputs = layer_module( 2025-09-07T09:02:14.0423335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0423682Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0424086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0424480Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0424878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0425262Z self_outputs = self.self( 2025-09-07T09:02:14.0425666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0426086Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0426554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0427119Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0427344Z 2025-09-07T09:02:14.0427419Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0427622Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0427841Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0428318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0428799Z layer_outputs = layer_module( 2025-09-07T09:02:14.0429128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0429464Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0429856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0430262Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0430641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0431024Z self_outputs = self.self( 2025-09-07T09:02:14.0431401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0431817Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0432283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0432824Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0433057Z 2025-09-07T09:02:14.0433134Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0433334Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0433559Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0434033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0434478Z layer_outputs = layer_module( 2025-09-07T09:02:14.0434804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0435149Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0435543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0435930Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0436311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0436697Z self_outputs = self.self( 2025-09-07T09:02:14.0437070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:02:14.0437452Z attn_scores += diagonal_mask 2025-09-07T09:02:14.0437567Z 2025-09-07T09:02:14.0437674Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0438148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0438600Z layer_outputs = layer_module( 2025-09-07T09:02:14.0461776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0462364Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0462790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0463256Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0463675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0464084Z self_outputs = self.self( 2025-09-07T09:02:14.0464481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:02:14.0464881Z attn_probs = nn.functional.softmax( 2025-09-07T09:02:14.0465021Z 2025-09-07T09:02:14.0465109Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0465356Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0465595Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0466099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0466584Z layer_outputs = layer_module( 2025-09-07T09:02:14.0466923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0467269Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0467664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0468053Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0468512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0468922Z self_outputs = self.self( 2025-09-07T09:02:14.0469322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0469764Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0470278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0470858Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:02:14.0471280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0471631Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0471782Z 2025-09-07T09:02:14.0471894Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0472394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0472865Z layer_outputs = layer_module( 2025-09-07T09:02:14.0473209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0473571Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0473974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0474372Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0474777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0475179Z self_outputs = self.self( 2025-09-07T09:02:14.0475567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0476057Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0476562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0477116Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:02:14.0477610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:02:14.0478061Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:02:14.0478419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0478763Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0478926Z 2025-09-07T09:02:14.0479033Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0479565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0480041Z layer_outputs = layer_module( 2025-09-07T09:02:14.0480384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0480760Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0481166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0481596Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0482006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0482405Z self_outputs = self.self( 2025-09-07T09:02:14.0482807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0483246Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0483741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0484313Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0484503Z 2025-09-07T09:02:14.0484613Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0485088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0485537Z layer_outputs = layer_module( 2025-09-07T09:02:14.0485866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0486210Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0486598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0486989Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0487381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0487771Z self_outputs = self.self( 2025-09-07T09:02:14.0488145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0488563Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0489048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0489572Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0489778Z 2025-09-07T09:02:14.0489887Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0490371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0490836Z layer_outputs = layer_module( 2025-09-07T09:02:14.0491168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0491512Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0491903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0492298Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0492683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0493085Z self_outputs = self.self( 2025-09-07T09:02:14.0493461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:02:14.0493957Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:02:14.0494640Z 2025-09-07T09:02:14.0494731Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0494936Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0495144Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0495348Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0495552Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0495747Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0495976Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0496476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0496940Z layer_outputs = layer_module( 2025-09-07T09:02:14.0497275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0497624Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0498013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0498409Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0498801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0499195Z self_outputs = self.self( 2025-09-07T09:02:14.0499575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:02:14.0499972Z query_vectors = self.query(hidden_states) 2025-09-07T09:02:14.0500113Z 2025-09-07T09:02:14.0500191Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0500393Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0500623Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0501102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0501562Z layer_outputs = layer_module( 2025-09-07T09:02:14.0501890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0502238Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0502634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0503022Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0503435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0503821Z self_outputs = self.self( 2025-09-07T09:02:14.0504193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0504640Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0505099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0505647Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0505883Z 2025-09-07T09:02:14.0505957Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0506157Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0506380Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0506867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0507320Z layer_outputs = layer_module( 2025-09-07T09:02:14.0507644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0507999Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0508386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0508764Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0509156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0509539Z self_outputs = self.self( 2025-09-07T09:02:14.0509909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0510316Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0510781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0511325Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0511554Z 2025-09-07T09:02:14.0511629Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0511828Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0512042Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0512524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0512971Z layer_outputs = layer_module( 2025-09-07T09:02:14.0513297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0513638Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0514022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0514414Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0514799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0515185Z self_outputs = self.self( 2025-09-07T09:02:14.0515554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:02:14.0515930Z attn_scores += diagonal_mask 2025-09-07T09:02:14.0516051Z 2025-09-07T09:02:14.0516148Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0516636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0517085Z layer_outputs = layer_module( 2025-09-07T09:02:14.0517404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0517761Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0518151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0518545Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0518935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0519313Z self_outputs = self.self( 2025-09-07T09:02:14.0519684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:02:14.0520090Z attn_probs = nn.functional.softmax( 2025-09-07T09:02:14.0520219Z 2025-09-07T09:02:14.0520301Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0520501Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0520717Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0521210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0521660Z layer_outputs = layer_module( 2025-09-07T09:02:14.0521986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0522313Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0522701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0523091Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0523476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0523859Z self_outputs = self.self( 2025-09-07T09:02:14.0524218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0524647Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0525133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0525678Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:02:14.0526070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0526392Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0526543Z 2025-09-07T09:02:14.0526643Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0527123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0527573Z layer_outputs = layer_module( 2025-09-07T09:02:14.0527899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0528227Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0528618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0529008Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0529389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0529765Z self_outputs = self.self( 2025-09-07T09:02:14.0530144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0530575Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0531072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0531599Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:02:14.0532073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:02:14.0532505Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:02:14.0532821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0533164Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0533309Z 2025-09-07T09:02:14.0533414Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0533893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0534353Z layer_outputs = layer_module( 2025-09-07T09:02:14.0534680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0535019Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0535405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0535782Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0536174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0536559Z self_outputs = self.self( 2025-09-07T09:02:14.0536937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0537361Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0537845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0538375Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0538575Z 2025-09-07T09:02:14.0538671Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0539160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0539613Z layer_outputs = layer_module( 2025-09-07T09:02:14.0539942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0540284Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0540671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0541060Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0541451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0541836Z self_outputs = self.self( 2025-09-07T09:02:14.0542207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0542633Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0543126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0543646Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0543842Z 2025-09-07T09:02:14.0543940Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0544416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0544882Z layer_outputs = layer_module( 2025-09-07T09:02:14.0545202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0545537Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0545924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0546314Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0546707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0547091Z self_outputs = self.self( 2025-09-07T09:02:14.0547458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:02:14.0547963Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:02:14.0548186Z 2025-09-07T09:02:14.0548269Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0548465Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0548664Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0548858Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0549051Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0549234Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0549453Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0549934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0550395Z layer_outputs = layer_module( 2025-09-07T09:02:14.0550731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0551070Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0551470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0551872Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0552270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0552661Z self_outputs = self.self( 2025-09-07T09:02:14.0553036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:02:14.0553443Z query_vectors = self.query(hidden_states) 2025-09-07T09:02:14.0553582Z 2025-09-07T09:02:14.0553656Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0553857Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0554078Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0554566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0555027Z layer_outputs = layer_module( 2025-09-07T09:02:14.0555359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0555704Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0556098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0556523Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0556921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0557311Z self_outputs = self.self( 2025-09-07T09:02:14.0557711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0558124Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0558599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0559159Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0559387Z 2025-09-07T09:02:14.0559470Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0559689Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0559913Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0560403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0560883Z layer_outputs = layer_module( 2025-09-07T09:02:14.0561215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0561549Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0561943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0562336Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0562732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0563125Z self_outputs = self.self( 2025-09-07T09:02:14.0563499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0563918Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0564388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0564942Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0565169Z 2025-09-07T09:02:14.0565252Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0565443Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0565671Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0566161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0566624Z layer_outputs = layer_module( 2025-09-07T09:02:14.0566958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0567296Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0567706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0568099Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0568523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0568911Z self_outputs = self.self( 2025-09-07T09:02:14.0569294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:02:14.0569700Z attn_scores += diagonal_mask 2025-09-07T09:02:14.0569816Z 2025-09-07T09:02:14.0569958Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0570433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0570872Z layer_outputs = layer_module( 2025-09-07T09:02:14.0571227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0571561Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0571954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0572341Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0572720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0573102Z self_outputs = self.self( 2025-09-07T09:02:14.0573497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:02:14.0573894Z attn_probs = nn.functional.softmax( 2025-09-07T09:02:14.0574016Z 2025-09-07T09:02:14.0574096Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0574313Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0574534Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0575010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0575458Z layer_outputs = layer_module( 2025-09-07T09:02:14.0575774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0576108Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0576503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0576890Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0577270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0577649Z self_outputs = self.self( 2025-09-07T09:02:14.0578018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0578439Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0578922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0579459Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:02:14.0579845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0580169Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0580318Z 2025-09-07T09:02:14.0580413Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0580888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0581338Z layer_outputs = layer_module( 2025-09-07T09:02:14.0581656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0581994Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0582380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0582767Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0583167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0583543Z self_outputs = self.self( 2025-09-07T09:02:14.0583911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0584346Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0584829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0585327Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:02:14.0585792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:02:14.0586225Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:02:14.0586555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0586885Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0587027Z 2025-09-07T09:02:14.0587123Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0587624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0588087Z layer_outputs = layer_module( 2025-09-07T09:02:14.0588421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0588767Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0589159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0589558Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0589953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0590347Z self_outputs = self.self( 2025-09-07T09:02:14.0590725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0591151Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0591648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0592177Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0592368Z 2025-09-07T09:02:14.0592473Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0592962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0593414Z layer_outputs = layer_module( 2025-09-07T09:02:14.0593741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0594080Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0594477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0594869Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0595257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0595646Z self_outputs = self.self( 2025-09-07T09:02:14.0596021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0596446Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0596960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0597471Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0597687Z 2025-09-07T09:02:14.0597784Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0598264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0598717Z layer_outputs = layer_module( 2025-09-07T09:02:14.0599041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0599369Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0599773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0600161Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0600545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0600955Z self_outputs = self.self( 2025-09-07T09:02:14.0601317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:02:14.0601808Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:02:14.0602037Z 2025-09-07T09:02:14.0602113Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0602313Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0602502Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0602691Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0602884Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0603076Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0603291Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0603773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0604229Z layer_outputs = layer_module( 2025-09-07T09:02:14.0604555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0604891Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0605274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0605660Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0606045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0606427Z self_outputs = self.self( 2025-09-07T09:02:14.0606794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:02:14.0607184Z query_vectors = self.query(hidden_states) 2025-09-07T09:02:14.0607321Z 2025-09-07T09:02:14.0607393Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0607589Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0607807Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0608279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0608725Z layer_outputs = layer_module( 2025-09-07T09:02:14.0609050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0609384Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0609791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0610171Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0610554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0610954Z self_outputs = self.self( 2025-09-07T09:02:14.0611321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0611734Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0612191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0612743Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0612978Z 2025-09-07T09:02:14.0613056Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0613253Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0613480Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0613957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0614422Z layer_outputs = layer_module( 2025-09-07T09:02:14.0614744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0615075Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0615459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0615846Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0616237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0616628Z self_outputs = self.self( 2025-09-07T09:02:14.0616996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0617404Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0617865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0618406Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0618629Z 2025-09-07T09:02:14.0618710Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0618908Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0619122Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0619604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0620054Z layer_outputs = layer_module( 2025-09-07T09:02:14.0620379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0620722Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0621102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0621486Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0621871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0622251Z self_outputs = self.self( 2025-09-07T09:02:14.0622629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:02:14.0623016Z attn_scores += diagonal_mask 2025-09-07T09:02:14.0623137Z 2025-09-07T09:02:14.0623235Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0623708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0624174Z layer_outputs = layer_module( 2025-09-07T09:02:14.0624490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0624820Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0625207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0625594Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0625997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0626370Z self_outputs = self.self( 2025-09-07T09:02:14.0626739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:02:14.0627140Z attn_probs = nn.functional.softmax( 2025-09-07T09:02:14.0627263Z 2025-09-07T09:02:14.0627344Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0627539Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0627749Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0628218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0628663Z layer_outputs = layer_module( 2025-09-07T09:02:14.0628985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0629313Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0629695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0630082Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0630464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0630840Z self_outputs = self.self( 2025-09-07T09:02:14.0631198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0631615Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0632096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0632637Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:02:14.0633027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0633344Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0633498Z 2025-09-07T09:02:14.0633593Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0634069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0634515Z layer_outputs = layer_module( 2025-09-07T09:02:14.0634834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0635162Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0635557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0635962Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0636351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0636735Z self_outputs = self.self( 2025-09-07T09:02:14.0637125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0637552Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0638037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0638543Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:02:14.0639030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:02:14.0639460Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:02:14.0639777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0640103Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0640270Z 2025-09-07T09:02:14.0640378Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0640848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0641298Z layer_outputs = layer_module( 2025-09-07T09:02:14.0641623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0641958Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0642347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0642729Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0643121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0643503Z self_outputs = self.self( 2025-09-07T09:02:14.0643877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0644300Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0644775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0645293Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0645487Z 2025-09-07T09:02:14.0645585Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0646059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0646505Z layer_outputs = layer_module( 2025-09-07T09:02:14.0646820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0647159Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0647546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0647927Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0648309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0648682Z self_outputs = self.self( 2025-09-07T09:02:14.0649067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0649485Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0649963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0650498Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0650686Z 2025-09-07T09:02:14.0650785Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0651264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0651716Z layer_outputs = layer_module( 2025-09-07T09:02:14.0652043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0652406Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0652790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0653182Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0653586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0653968Z self_outputs = self.self( 2025-09-07T09:02:14.0654341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:02:14.0654825Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:02:14.0655058Z 2025-09-07T09:02:14.0655135Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0655338Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0655535Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0655722Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0655914Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0656105Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0656323Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0656795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0657242Z layer_outputs = layer_module( 2025-09-07T09:02:14.0657563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0657898Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0658291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0658670Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0659060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0659441Z self_outputs = self.self( 2025-09-07T09:02:14.0659804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:02:14.0660199Z query_vectors = self.query(hidden_states) 2025-09-07T09:02:14.0660328Z 2025-09-07T09:02:14.0660400Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0660594Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0660812Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0661287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0661728Z layer_outputs = layer_module( 2025-09-07T09:02:14.0662068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0662404Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0662792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0663193Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0663573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0663956Z self_outputs = self.self( 2025-09-07T09:02:14.0664326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0664736Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0665218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0665755Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0665985Z 2025-09-07T09:02:14.0666057Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0666254Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0666491Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0666962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0667400Z layer_outputs = layer_module( 2025-09-07T09:02:14.0667723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0668063Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0668488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0668877Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0669273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0669666Z self_outputs = self.self( 2025-09-07T09:02:14.0670051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0670484Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0670942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0671497Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0671734Z 2025-09-07T09:02:14.0671810Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0672018Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0672248Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0672731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0673198Z layer_outputs = layer_module( 2025-09-07T09:02:14.0673528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0673870Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0674267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0674654Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0675045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0675473Z self_outputs = self.self( 2025-09-07T09:02:14.0675856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:02:14.0676244Z attn_scores += diagonal_mask 2025-09-07T09:02:14.0676369Z 2025-09-07T09:02:14.0676472Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0676982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0677439Z layer_outputs = layer_module( 2025-09-07T09:02:14.0677768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0678104Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0678520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0678943Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0679342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0679731Z self_outputs = self.self( 2025-09-07T09:02:14.0680137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:02:14.0680543Z attn_probs = nn.functional.softmax( 2025-09-07T09:02:14.0680676Z 2025-09-07T09:02:14.0680749Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0680952Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0681176Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0681661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0682125Z layer_outputs = layer_module( 2025-09-07T09:02:14.0682463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0682802Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0683189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0683581Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0683964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0684348Z self_outputs = self.self( 2025-09-07T09:02:14.0684719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0685136Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0685623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0686164Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:02:14.0686562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0686895Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0687037Z 2025-09-07T09:02:14.0687133Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0687614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0688063Z layer_outputs = layer_module( 2025-09-07T09:02:14.0688388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0688731Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0689134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0689522Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0689914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0690309Z self_outputs = self.self( 2025-09-07T09:02:14.0690676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0691092Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0691577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0692095Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:02:14.0692573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:02:14.0693011Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:02:14.0693323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0693674Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0693826Z 2025-09-07T09:02:14.0693926Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0694410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0694864Z layer_outputs = layer_module( 2025-09-07T09:02:14.0695189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0695535Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0695934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0696331Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0696722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0697115Z self_outputs = self.self( 2025-09-07T09:02:14.0697494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0697924Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0698419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0698946Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0699149Z 2025-09-07T09:02:14.0699251Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0699732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0700186Z layer_outputs = layer_module( 2025-09-07T09:02:14.0700515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0700848Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0701245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0701636Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0702031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0702432Z self_outputs = self.self( 2025-09-07T09:02:14.0702801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0703224Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0703743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0704259Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0704446Z 2025-09-07T09:02:14.0704554Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0705021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0705469Z layer_outputs = layer_module( 2025-09-07T09:02:14.0705826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0706166Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0706555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0706951Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0707342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0707721Z self_outputs = self.self( 2025-09-07T09:02:14.0708085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:02:14.0708570Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:02:14.0708789Z 2025-09-07T09:02:14.0708867Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0709056Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0709246Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0709440Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0709624Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0709816Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0710039Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0710511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0710956Z layer_outputs = layer_module( 2025-09-07T09:02:14.0711268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0711607Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0711995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0712378Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0712750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0713128Z self_outputs = self.self( 2025-09-07T09:02:14.0713494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:02:14.0713880Z query_vectors = self.query(hidden_states) 2025-09-07T09:02:14.0714007Z 2025-09-07T09:02:14.0714081Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0714264Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0714478Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0714965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0715411Z layer_outputs = layer_module( 2025-09-07T09:02:14.0715724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0716056Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0716451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0716828Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0717206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0717579Z self_outputs = self.self( 2025-09-07T09:02:14.0717943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0718354Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0718817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0719361Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0719600Z 2025-09-07T09:02:14.0719670Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0719859Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0720074Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0720544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0720993Z layer_outputs = layer_module( 2025-09-07T09:02:14.0721310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0721638Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0722027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0722415Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0722791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0723174Z self_outputs = self.self( 2025-09-07T09:02:14.0723538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0723942Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0724400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0724932Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0725163Z 2025-09-07T09:02:14.0725238Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0725433Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0725651Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0726123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0726565Z layer_outputs = layer_module( 2025-09-07T09:02:14.0726887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0727222Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0727603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0727989Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0728377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0728756Z self_outputs = self.self( 2025-09-07T09:02:14.0729126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:02:14.0729521Z attn_scores += diagonal_mask 2025-09-07T09:02:14.0729632Z 2025-09-07T09:02:14.0729735Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0730199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0730643Z layer_outputs = layer_module( 2025-09-07T09:02:14.0730965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0731298Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0731694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0732082Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0732466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0732859Z self_outputs = self.self( 2025-09-07T09:02:14.0733220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:02:14.0733603Z attn_probs = nn.functional.softmax( 2025-09-07T09:02:14.0733731Z 2025-09-07T09:02:14.0733805Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0734001Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0734220Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0734694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0735133Z layer_outputs = layer_module( 2025-09-07T09:02:14.0735452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0735783Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0736172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0736548Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0736927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0737302Z self_outputs = self.self( 2025-09-07T09:02:14.0737664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0738090Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0738565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0739097Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:02:14.0739494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0739818Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0739960Z 2025-09-07T09:02:14.0740066Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0740537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0740985Z layer_outputs = layer_module( 2025-09-07T09:02:14.0741319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0741659Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0742053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0742453Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0742843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0743228Z self_outputs = self.self( 2025-09-07T09:02:14.0743600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0744024Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0744524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0745026Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:02:14.0745495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:02:14.0745949Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:02:14.0746266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0746583Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0746729Z 2025-09-07T09:02:14.0746827Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0747306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0747749Z layer_outputs = layer_module( 2025-09-07T09:02:14.0748073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0748401Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0748797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0749187Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0749565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0749942Z self_outputs = self.self( 2025-09-07T09:02:14.0750310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0750724Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0751214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0751731Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0751923Z 2025-09-07T09:02:14.0752022Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0752502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0752949Z layer_outputs = layer_module( 2025-09-07T09:02:14.0753278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0753625Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0754022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0754422Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0754843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0755245Z self_outputs = self.self( 2025-09-07T09:02:14.0755634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0756077Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0756574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0757110Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0757303Z 2025-09-07T09:02:14.0757410Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0757918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0758373Z layer_outputs = layer_module( 2025-09-07T09:02:14.0758705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0759076Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0759484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0759890Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0760287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0760683Z self_outputs = self.self( 2025-09-07T09:02:14.0761072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:02:14.0761583Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:02:14.0761813Z 2025-09-07T09:02:14.0761898Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0762103Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0762309Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0762518Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0762723Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0762920Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0763152Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0763653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0764123Z layer_outputs = layer_module( 2025-09-07T09:02:14.0764458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0764817Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0765220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0765624Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0766026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0766418Z self_outputs = self.self( 2025-09-07T09:02:14.0766801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:02:14.0767213Z query_vectors = self.query(hidden_states) 2025-09-07T09:02:14.0767348Z 2025-09-07T09:02:14.0767434Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0767640Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0767865Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0768439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0768896Z layer_outputs = layer_module( 2025-09-07T09:02:14.0769223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0769587Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0769977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0770363Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0770754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0771139Z self_outputs = self.self( 2025-09-07T09:02:14.0771525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0771942Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0772407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0772987Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0773214Z 2025-09-07T09:02:14.0773295Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0773487Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0773711Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0774193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0774650Z layer_outputs = layer_module( 2025-09-07T09:02:14.0774988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0775326Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0775721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0776116Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0776509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0776896Z self_outputs = self.self( 2025-09-07T09:02:14.0777262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0777681Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0778148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0778694Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0778917Z 2025-09-07T09:02:14.0778999Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0779192Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0779416Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0779896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0780345Z layer_outputs = layer_module( 2025-09-07T09:02:14.0780664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0781002Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0781409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0781801Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0782190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0782564Z self_outputs = self.self( 2025-09-07T09:02:14.0782959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:02:14.0783353Z attn_scores += diagonal_mask 2025-09-07T09:02:14.0783469Z 2025-09-07T09:02:14.0783578Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0784069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0784520Z layer_outputs = layer_module( 2025-09-07T09:02:14.0784866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0785206Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0785596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0785992Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0786380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0786755Z self_outputs = self.self( 2025-09-07T09:02:14.0787122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:02:14.0787509Z attn_probs = nn.functional.softmax( 2025-09-07T09:02:14.0787633Z 2025-09-07T09:02:14.0787707Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0787899Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0788123Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0788600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0789050Z layer_outputs = layer_module( 2025-09-07T09:02:14.0789376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0789711Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0790100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0790488Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0790869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0790943Z self_outputs = self.self( 2025-09-07T09:02:14.0791209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0791329Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0791652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0791829Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:02:14.0792009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0792103Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0792106Z 2025-09-07T09:02:14.0792212Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0792552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0792629Z layer_outputs = layer_module( 2025-09-07T09:02:14.0792833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0792916Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0793194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0793264Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0793528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0793593Z self_outputs = self.self( 2025-09-07T09:02:14.0793861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0793982Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0794310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0794442Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:02:14.0794760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:02:14.0794854Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:02:14.0795033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0795135Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0795138Z 2025-09-07T09:02:14.0795234Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0795564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0795640Z layer_outputs = layer_module( 2025-09-07T09:02:14.0795848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0795930Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0796192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0796271Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0796533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0796600Z self_outputs = self.self( 2025-09-07T09:02:14.0796869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0796980Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0797312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0797456Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0797462Z 2025-09-07T09:02:14.0797559Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0797892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0797960Z layer_outputs = layer_module( 2025-09-07T09:02:14.0798174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0798247Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0798530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0798602Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0798866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0798955Z self_outputs = self.self( 2025-09-07T09:02:14.0799214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0799330Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0799660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0799805Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0799808Z 2025-09-07T09:02:14.0799924Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0800246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0800321Z layer_outputs = layer_module( 2025-09-07T09:02:14.0800541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0800622Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0800881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0800952Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0801221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0801286Z self_outputs = self.self( 2025-09-07T09:02:14.0801551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:02:14.0801724Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:02:14.0801727Z 2025-09-07T09:02:14.0801807Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0801884Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0801954Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0802034Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0802103Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0802179Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0802277Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0802606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0802684Z layer_outputs = layer_module( 2025-09-07T09:02:14.0802889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0802972Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0803232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0803306Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0803576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0803641Z self_outputs = self.self( 2025-09-07T09:02:14.0803904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:02:14.0803986Z query_vectors = self.query(hidden_states) 2025-09-07T09:02:14.0803989Z 2025-09-07T09:02:14.0804066Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0804154Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0804253Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0804582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0804666Z layer_outputs = layer_module( 2025-09-07T09:02:14.0804886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0804960Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0805224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0805304Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0805564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0805651Z self_outputs = self.self( 2025-09-07T09:02:14.0805913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0806019Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0806353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0806526Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0806529Z 2025-09-07T09:02:14.0806609Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0806681Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0806787Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0807114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0807180Z layer_outputs = layer_module( 2025-09-07T09:02:14.0807395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0807467Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0807739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0807809Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0808077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0808141Z self_outputs = self.self( 2025-09-07T09:02:14.0808399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:02:14.0808505Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:02:14.0808822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:02:14.0808998Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:02:14.0809004Z 2025-09-07T09:02:14.0809077Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0809150Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0809253Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0809578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0809651Z layer_outputs = layer_module( 2025-09-07T09:02:14.0809860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0809957Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0810217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0810286Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0810551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0810633Z self_outputs = self.self( 2025-09-07T09:02:14.0810902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:02:14.0810973Z attn_scores += diagonal_mask 2025-09-07T09:02:14.0810976Z 2025-09-07T09:02:14.0811073Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0811423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0811492Z layer_outputs = layer_module( 2025-09-07T09:02:14.0811709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0811781Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0812066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0812141Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0812397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0812470Z self_outputs = self.self( 2025-09-07T09:02:14.0812726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:02:14.0812808Z attn_probs = nn.functional.softmax( 2025-09-07T09:02:14.0812811Z 2025-09-07T09:02:14.0812885Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0812957Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0813062Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0813389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0813463Z layer_outputs = layer_module( 2025-09-07T09:02:14.0813667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0813748Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0814006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0814074Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0814342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0814407Z self_outputs = self.self( 2025-09-07T09:02:14.0814668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0814778Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0815104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0815270Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:02:14.0815449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0815548Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0815551Z 2025-09-07T09:02:14.0815647Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0815994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0816062Z layer_outputs = layer_module( 2025-09-07T09:02:14.0816266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0816365Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0816626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0816705Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0816963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0817026Z self_outputs = self.self( 2025-09-07T09:02:14.0817314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0817420Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0817755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0817898Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:02:14.0818201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:02:14.0818288Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:02:14.0818465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:02:14.0818562Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:02:14.0818565Z 2025-09-07T09:02:14.0818663Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0818997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0819064Z layer_outputs = layer_module( 2025-09-07T09:02:14.0819280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0819354Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0819617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0819699Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0819958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0820032Z self_outputs = self.self( 2025-09-07T09:02:14.0820289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0820397Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0820727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0820869Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0820872Z 2025-09-07T09:02:14.0820977Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0821300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0821377Z layer_outputs = layer_module( 2025-09-07T09:02:14.0821584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0821671Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0821944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0822017Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0822303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0822368Z self_outputs = self.self( 2025-09-07T09:02:14.0822631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:02:14.0822736Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:02:14.0823059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:02:14.0823219Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:02:14.0823222Z 2025-09-07T09:02:14.0823319Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:14.0823652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:02:14.0823735Z layer_outputs = layer_module( 2025-09-07T09:02:14.0823946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:02:14.0824019Z return super().__call__(*args, **kwargs) 2025-09-07T09:02:14.0824278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:02:14.0824356Z self_attn_outputs = self.attention( 2025-09-07T09:02:14.0824617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:02:14.0824689Z self_outputs = self.self( 2025-09-07T09:02:14.0824947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:02:14.0825116Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:02:14.0825122Z 2025-09-07T09:02:14.0825205Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0825278Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0825359Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0825432Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0825503Z cudagraph partition due to non gpu ops 2025-09-07T09:02:14.0825581Z cudagraph partition due to non gpu ops 2025-09-07T09:02:28.0972746Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:28.0973385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1716, in torch_dynamo_resume_in_forward_at_1703 2025-09-07T09:02:28.0973891Z prediction_scores = self.lm_head(sequence_output) 2025-09-07T09:02:28.0974323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1333, in forward 2025-09-07T09:02:28.0974735Z x = self.dense(features) 2025-09-07T09:02:28.0974860Z 2025-09-07T09:02:28.0974945Z cudagraph partition due to non gpu ops 2025-09-07T09:02:28.0975155Z cudagraph partition due to non gpu ops 2025-09-07T09:02:28.0975354Z cudagraph partition due to non gpu ops 2025-09-07T09:02:28.0975544Z cudagraph partition due to non gpu ops 2025-09-07T09:02:29.4902564Z 2025-09-07T09:02:29.6108431Z running benchmark: 0% 0/30 [00:00 2025-09-07T09:02:49.2087502Z torch.index_select(position_encodings[i], 0, position_ids[i]).unsqueeze(0) 2025-09-07T09:02:49.2087776Z 2025-09-07T09:02:49.8132141Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:49.8132743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 370, in torch_dynamo_resume_in_forward_at_369 2025-09-07T09:02:49.8133236Z embeddings = embeddings + position_embeddings 2025-09-07T09:02:49.8133387Z 2025-09-07T09:02:54.4236431Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4236778Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4237356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4237840Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4238262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4238663Z layer_outputs = layer( 2025-09-07T09:02:54.4239040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4239427Z attn_outputs = self.attention( 2025-09-07T09:02:54.4239808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4240213Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4240621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:02:54.4241124Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:02:54.4241706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:02:54.4242224Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T09:02:54.4242409Z 2025-09-07T09:02:54.4242514Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4243004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4243489Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4243890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4244263Z layer_outputs = layer( 2025-09-07T09:02:54.4244938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4245338Z attn_outputs = self.attention( 2025-09-07T09:02:54.4245718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4246177Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4246560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:02:54.4247044Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:02:54.4247616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:02:54.4248118Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T09:02:54.4248334Z 2025-09-07T09:02:54.4248445Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4249067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4249566Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4249960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4250339Z layer_outputs = layer( 2025-09-07T09:02:54.4250699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4251075Z attn_outputs = self.attention( 2025-09-07T09:02:54.4251443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4251837Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4252236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:02:54.4252713Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:02:54.4253275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:02:54.4253783Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T09:02:54.4253955Z 2025-09-07T09:02:54.4254033Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4254241Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4254437Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4254629Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4254821Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4255012Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4255234Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4255698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4256161Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4256554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4256919Z layer_outputs = layer( 2025-09-07T09:02:54.4257292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4257670Z attn_outputs = self.attention( 2025-09-07T09:02:54.4258040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4258486Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4258873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T09:02:54.4259351Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4259915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4260389Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4260585Z 2025-09-07T09:02:54.4260671Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4260868Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4261095Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4261588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4262049Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4262434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4262822Z layer_outputs = layer( 2025-09-07T09:02:54.4263186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4263559Z attn_outputs = self.attention( 2025-09-07T09:02:54.4263927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4264310Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4264703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T09:02:54.4265174Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4265662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:02:54.4266052Z return torch.cat(slices, dim=3) 2025-09-07T09:02:54.4266176Z 2025-09-07T09:02:54.4266274Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4266741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4267195Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4267583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4267958Z layer_outputs = layer( 2025-09-07T09:02:54.4268403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4268779Z attn_outputs = self.attention( 2025-09-07T09:02:54.4269153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4269554Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4269960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T09:02:54.4270419Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:02:54.4270627Z 2025-09-07T09:02:54.4270730Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4271210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4271680Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4272115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4272493Z layer_outputs = layer( 2025-09-07T09:02:54.4272861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4273271Z attn_outputs = self.attention( 2025-09-07T09:02:54.4273653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4274044Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4274447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T09:02:54.4274904Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:02:54.4275105Z 2025-09-07T09:02:54.4275207Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4275424Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4275621Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4275826Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4276054Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4276558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4277027Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4277419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4277805Z layer_outputs = layer( 2025-09-07T09:02:54.4278174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4278560Z attn_outputs = self.attention( 2025-09-07T09:02:54.4278934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4279342Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4279744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T09:02:54.4280248Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4280766Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4281234Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4281442Z 2025-09-07T09:02:54.4281518Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4281719Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4281949Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4282429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4282890Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4283301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4283688Z layer_outputs = layer( 2025-09-07T09:02:54.4284054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4284442Z attn_outputs = self.attention( 2025-09-07T09:02:54.4284818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4285223Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4285646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T09:02:54.4286074Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:02:54.4286234Z 2025-09-07T09:02:54.4286343Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4286832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4287301Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4287701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4288079Z layer_outputs = layer( 2025-09-07T09:02:54.4288437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4288843Z attn_outputs = self.attention( 2025-09-07T09:02:54.4289228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4289629Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4290030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T09:02:54.4290473Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:02:54.4290643Z 2025-09-07T09:02:54.4290745Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4291228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4291694Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4292094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4292468Z layer_outputs = layer( 2025-09-07T09:02:54.4292830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4293212Z attn_outputs = self.attention( 2025-09-07T09:02:54.4293590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4293988Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4294378Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T09:02:54.4294891Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T09:02:54.4295439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T09:02:54.4295932Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T09:02:54.4296120Z 2025-09-07T09:02:54.4296205Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4296407Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4296616Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4296814Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4297015Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4297205Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4297404Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4297599Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4297823Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4298289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4298753Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4299167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4299552Z layer_outputs = layer( 2025-09-07T09:02:54.4299913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4300310Z attn_outputs = self.attention( 2025-09-07T09:02:54.4300691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4301089Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4301486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:02:54.4301932Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:02:54.4302410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T09:02:54.4302889Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T09:02:54.4303095Z 2025-09-07T09:02:54.4303198Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4303714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4304180Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4304572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4304959Z layer_outputs = layer( 2025-09-07T09:02:54.4305332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4305725Z attn_outputs = self.attention( 2025-09-07T09:02:54.4306106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4306501Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4306905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:02:54.4307358Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:02:54.4307822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T09:02:54.4308288Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T09:02:54.4308483Z 2025-09-07T09:02:54.4308584Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4309068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4309534Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4309933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4310317Z layer_outputs = layer( 2025-09-07T09:02:54.4310672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4311057Z attn_outputs = self.attention( 2025-09-07T09:02:54.4311431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4311835Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4312228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:02:54.4313627Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:02:54.4314091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T09:02:54.4314511Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T09:02:54.4314677Z 2025-09-07T09:02:54.4314787Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4315261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4315742Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4316147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4316534Z layer_outputs = layer( 2025-09-07T09:02:54.4316920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4317300Z attn_outputs = self.attention( 2025-09-07T09:02:54.4317678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4318099Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4318496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:02:54.4318987Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:02:54.4319553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:02:54.4320055Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T09:02:54.4320493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T09:02:54.4320976Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T09:02:54.4321175Z 2025-09-07T09:02:54.4321263Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4321496Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4321978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4322442Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4322845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4323225Z layer_outputs = layer( 2025-09-07T09:02:54.4323586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4323972Z attn_outputs = self.attention( 2025-09-07T09:02:54.4324353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4324760Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4325161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:02:54.4325644Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:02:54.4326223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:02:54.4326721Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T09:02:54.4327166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T09:02:54.4327589Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T09:02:54.4327732Z 2025-09-07T09:02:54.4327839Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4328053Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4328303Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4328780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4329242Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4329634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4330016Z layer_outputs = layer( 2025-09-07T09:02:54.4330401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4330788Z attn_outputs = self.attention( 2025-09-07T09:02:54.4331160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4331563Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4331979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4332391Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4332816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T09:02:54.4333327Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4333855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4334330Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4334524Z 2025-09-07T09:02:54.4334608Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4334814Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4335040Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4335519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4335990Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4336390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4336771Z layer_outputs = layer( 2025-09-07T09:02:54.4337132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4337525Z attn_outputs = self.attention( 2025-09-07T09:02:54.4337902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4338300Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4338695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4339103Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4339505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T09:02:54.4340010Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4340528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:02:54.4340946Z return torch.cat(slices, dim=3) 2025-09-07T09:02:54.4341080Z 2025-09-07T09:02:54.4341155Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4341352Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4341547Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4341785Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4342251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4342707Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4343102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4343472Z layer_outputs = layer( 2025-09-07T09:02:54.4343843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4344214Z attn_outputs = self.attention( 2025-09-07T09:02:54.4344581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4344976Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4345383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4345775Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4346169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T09:02:54.4346640Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4347136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4347597Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4347787Z 2025-09-07T09:02:54.4347870Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4348088Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4348556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4349012Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4349404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4349771Z layer_outputs = layer( 2025-09-07T09:02:54.4350129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4350502Z attn_outputs = self.attention( 2025-09-07T09:02:54.4350881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4351270Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4351650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4352063Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4352458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T09:02:54.4352927Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4353415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:02:54.4353797Z return torch.cat(slices, dim=3) 2025-09-07T09:02:54.4353926Z 2025-09-07T09:02:54.4354038Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4354507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4354967Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4355378Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4355745Z layer_outputs = layer( 2025-09-07T09:02:54.4356105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4356477Z attn_outputs = self.attention( 2025-09-07T09:02:54.4356848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4357267Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4357653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T09:02:54.4358150Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T09:02:54.4358696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T09:02:54.4359131Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T09:02:54.4359272Z 2025-09-07T09:02:54.4359380Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4359843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4360302Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4360701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4361081Z layer_outputs = layer( 2025-09-07T09:02:54.4361440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4361823Z attn_outputs = self.attention( 2025-09-07T09:02:54.4362200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4362594Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4362985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4363379Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4363787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T09:02:54.4364240Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:02:54.4364431Z 2025-09-07T09:02:54.4364538Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4365007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4365457Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4365847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4366221Z layer_outputs = layer( 2025-09-07T09:02:54.4366578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4366957Z attn_outputs = self.attention( 2025-09-07T09:02:54.4367338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4367732Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4368119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4368573Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4368975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T09:02:54.4369447Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4369946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4370407Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4370598Z 2025-09-07T09:02:54.4370724Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4370952Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4371414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4371905Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4372300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4372671Z layer_outputs = layer( 2025-09-07T09:02:54.4373026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4373397Z attn_outputs = self.attention( 2025-09-07T09:02:54.4373778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4374180Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4374569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4374969Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4375364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T09:02:54.4375777Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:02:54.4375942Z 2025-09-07T09:02:54.4376040Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4376509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4376971Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4377357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4377730Z layer_outputs = layer( 2025-09-07T09:02:54.4378088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4378470Z attn_outputs = self.attention( 2025-09-07T09:02:54.4378835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4379237Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4379632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4380034Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4380433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T09:02:54.4380868Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:02:54.4381033Z 2025-09-07T09:02:54.4381111Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4381339Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4381803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4382288Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4382676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4383052Z layer_outputs = layer( 2025-09-07T09:02:54.4383412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4383789Z attn_outputs = self.attention( 2025-09-07T09:02:54.4384176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4384566Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4384955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T09:02:54.4385470Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T09:02:54.4386008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T09:02:54.4386485Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T09:02:54.4386666Z 2025-09-07T09:02:54.4386740Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4386943Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4387143Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4387341Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4387528Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4387724Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4387918Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4388117Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4388302Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4388494Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4388688Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4388881Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4389091Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4389552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4390006Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4390398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4390768Z layer_outputs = layer( 2025-09-07T09:02:54.4391117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4391493Z attn_outputs = self.attention( 2025-09-07T09:02:54.4391860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4392253Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4392644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T09:02:54.4393107Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4393614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4394074Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4394263Z 2025-09-07T09:02:54.4394347Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4394551Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4394787Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4395253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4395712Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4396104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4396471Z layer_outputs = layer( 2025-09-07T09:02:54.4396847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4397229Z attn_outputs = self.attention( 2025-09-07T09:02:54.4397606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4398013Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4398395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T09:02:54.4398864Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4399348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:02:54.4399738Z return torch.cat(slices, dim=3) 2025-09-07T09:02:54.4399859Z 2025-09-07T09:02:54.4399965Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4400424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4400878Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4401268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4401639Z layer_outputs = layer( 2025-09-07T09:02:54.4401994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4402359Z attn_outputs = self.attention( 2025-09-07T09:02:54.4402726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4403114Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4403502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T09:02:54.4403944Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:02:54.4404133Z 2025-09-07T09:02:54.4404229Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4404694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4405147Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4405530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4405891Z layer_outputs = layer( 2025-09-07T09:02:54.4406235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4406607Z attn_outputs = self.attention( 2025-09-07T09:02:54.4406992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4407387Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4407770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T09:02:54.4408237Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:02:54.4408431Z 2025-09-07T09:02:54.4408506Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4408708Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4408905Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4409095Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4409318Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4409800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4410261Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4410650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4411034Z layer_outputs = layer( 2025-09-07T09:02:54.4411393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4411767Z attn_outputs = self.attention( 2025-09-07T09:02:54.4412137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4412521Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4412913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T09:02:54.4413396Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4413894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4414359Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4414553Z 2025-09-07T09:02:54.4414630Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4414831Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4415053Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4415521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4415976Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4416361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4416734Z layer_outputs = layer( 2025-09-07T09:02:54.4417086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4417460Z attn_outputs = self.attention( 2025-09-07T09:02:54.4417830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4418217Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4418609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T09:02:54.4419017Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:02:54.4419170Z 2025-09-07T09:02:54.4419278Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4419757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4420210Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4420603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4420987Z layer_outputs = layer( 2025-09-07T09:02:54.4421345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4421707Z attn_outputs = self.attention( 2025-09-07T09:02:54.4422074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4422465Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4422856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T09:02:54.4423279Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:02:54.4423436Z 2025-09-07T09:02:54.4423532Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4423999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4424476Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4424872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4425241Z layer_outputs = layer( 2025-09-07T09:02:54.4425597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4425982Z attn_outputs = self.attention( 2025-09-07T09:02:54.4426359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4426757Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4427154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T09:02:54.4427652Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T09:02:54.4428206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T09:02:54.4428699Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T09:02:54.4428884Z 2025-09-07T09:02:54.4428968Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4429177Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4429370Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4429571Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4429773Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4429967Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4430156Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4430353Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4430578Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4431060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4431512Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4431914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4432290Z layer_outputs = layer( 2025-09-07T09:02:54.4432654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4433036Z attn_outputs = self.attention( 2025-09-07T09:02:54.4433422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4433818Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4434210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:02:54.4434666Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:02:54.4435119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T09:02:54.4435572Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T09:02:54.4435770Z 2025-09-07T09:02:54.4435868Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4436347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4436803Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4437191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4437576Z layer_outputs = layer( 2025-09-07T09:02:54.4437932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4438307Z attn_outputs = self.attention( 2025-09-07T09:02:54.4438678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4439071Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4439455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:02:54.4439897Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:02:54.4440353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T09:02:54.4440805Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T09:02:54.4440991Z 2025-09-07T09:02:54.4441099Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4441560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4442017Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4442405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4442778Z layer_outputs = layer( 2025-09-07T09:02:54.4443137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4443502Z attn_outputs = self.attention( 2025-09-07T09:02:54.4443875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4444271Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4444667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:02:54.4445099Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:02:54.4445552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T09:02:54.4445965Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T09:02:54.4446113Z 2025-09-07T09:02:54.4446212Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4446696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4447153Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4447538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4447932Z layer_outputs = layer( 2025-09-07T09:02:54.4448294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4448673Z attn_outputs = self.attention( 2025-09-07T09:02:54.4449034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4449431Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4449837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:02:54.4450318Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:02:54.4450883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:02:54.4451388Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T09:02:54.4451814Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T09:02:54.4452285Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T09:02:54.4452485Z 2025-09-07T09:02:54.4452568Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4452797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4453261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4453722Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4454118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4454496Z layer_outputs = layer( 2025-09-07T09:02:54.4454856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4455224Z attn_outputs = self.attention( 2025-09-07T09:02:54.4455599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4455996Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4456393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:02:54.4456873Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:02:54.4457433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:02:54.4457926Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T09:02:54.4458352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T09:02:54.4458765Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T09:02:54.4458903Z 2025-09-07T09:02:54.4458986Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4459181Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4459404Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4459908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4460364Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4460742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4461134Z layer_outputs = layer( 2025-09-07T09:02:54.4461491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4461865Z attn_outputs = self.attention( 2025-09-07T09:02:54.4462237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4462621Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4463032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4463436Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4463837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T09:02:54.4464354Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4464859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4465315Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4465514Z 2025-09-07T09:02:54.4465588Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4465793Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4466019Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4466482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4466941Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4467342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4467721Z layer_outputs = layer( 2025-09-07T09:02:54.4468075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4468479Z attn_outputs = self.attention( 2025-09-07T09:02:54.4468860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4469254Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4469657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4470067Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4470465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T09:02:54.4470963Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4471483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:02:54.4471880Z return torch.cat(slices, dim=3) 2025-09-07T09:02:54.4472003Z 2025-09-07T09:02:54.4472088Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4472287Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4472483Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4472708Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4473236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4473690Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4474080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4474474Z layer_outputs = layer( 2025-09-07T09:02:54.4474827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4475204Z attn_outputs = self.attention( 2025-09-07T09:02:54.4475568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4475962Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4476372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4476774Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4477167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T09:02:54.4477665Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4478160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4478621Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4478811Z 2025-09-07T09:02:54.4478896Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4479111Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4479583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4480047Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4480441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4480822Z layer_outputs = layer( 2025-09-07T09:02:54.4481180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4481563Z attn_outputs = self.attention( 2025-09-07T09:02:54.4481938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4482335Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4482727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4483124Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4483530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T09:02:54.4484006Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4484504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:02:54.4484899Z return torch.cat(slices, dim=3) 2025-09-07T09:02:54.4485023Z 2025-09-07T09:02:54.4485123Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4485591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4486048Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4486460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4486832Z layer_outputs = layer( 2025-09-07T09:02:54.4487185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4487586Z attn_outputs = self.attention( 2025-09-07T09:02:54.4487956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4488351Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4488742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T09:02:54.4489231Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T09:02:54.4489909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T09:02:54.4490390Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T09:02:54.4490529Z 2025-09-07T09:02:54.4490639Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4491107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4491576Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4491967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4492339Z layer_outputs = layer( 2025-09-07T09:02:54.4492691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4493064Z attn_outputs = self.attention( 2025-09-07T09:02:54.4493428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4493822Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4494210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4494613Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4495002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T09:02:54.4495445Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:02:54.4495639Z 2025-09-07T09:02:54.4495741Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4496207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4496665Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4497047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4497415Z layer_outputs = layer( 2025-09-07T09:02:54.4497772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4498149Z attn_outputs = self.attention( 2025-09-07T09:02:54.4498516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4498899Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4499288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4499687Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4500102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T09:02:54.4500586Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4501077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4501557Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4501754Z 2025-09-07T09:02:54.4501830Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4502060Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4502528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4502975Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4503389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4503762Z layer_outputs = layer( 2025-09-07T09:02:54.4504115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4504520Z attn_outputs = self.attention( 2025-09-07T09:02:54.4504888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4505286Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4505671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4506075Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4506477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T09:02:54.4506884Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:02:54.4507051Z 2025-09-07T09:02:54.4507147Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4507615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4508076Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4508467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4508829Z layer_outputs = layer( 2025-09-07T09:02:54.4509186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4509562Z attn_outputs = self.attention( 2025-09-07T09:02:54.4509936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4510320Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4510709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4511107Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4511502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T09:02:54.4511907Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:02:54.4512063Z 2025-09-07T09:02:54.4512139Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4512362Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4512830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4513309Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4513702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4514070Z layer_outputs = layer( 2025-09-07T09:02:54.4514448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4514827Z attn_outputs = self.attention( 2025-09-07T09:02:54.4515202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4515600Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4515985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T09:02:54.4516501Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T09:02:54.4517040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T09:02:54.4517523Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T09:02:54.4517723Z 2025-09-07T09:02:54.4517811Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4518012Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4518243Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4518716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4519176Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4519566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4519949Z layer_outputs = layer( 2025-09-07T09:02:54.4520313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1601, in forward 2025-09-07T09:02:54.4520713Z attn_output = prev_attn_output + attn_output 2025-09-07T09:02:54.4520855Z 2025-09-07T09:02:54.4520938Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4521133Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4521334Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4521530Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4521757Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4522222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4522687Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4523092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4523475Z layer_outputs = layer( 2025-09-07T09:02:54.4523835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:02:54.4524257Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:02:54.4524435Z 2025-09-07T09:02:54.4524512Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4524785Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4524984Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4525182Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4525373Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4525571Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4525796Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4526285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4526739Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4527131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4527520Z layer_outputs = layer( 2025-09-07T09:02:54.4527875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4528246Z attn_outputs = self.attention( 2025-09-07T09:02:54.4528608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4528996Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4529385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T09:02:54.4529875Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4530367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4530839Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4531040Z 2025-09-07T09:02:54.4531112Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4531312Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4531537Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4531997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4532457Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4532858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4533241Z layer_outputs = layer( 2025-09-07T09:02:54.4533598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4533974Z attn_outputs = self.attention( 2025-09-07T09:02:54.4534357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4534756Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4535153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T09:02:54.4535631Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4536116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:02:54.4536514Z return torch.cat(slices, dim=3) 2025-09-07T09:02:54.4536652Z 2025-09-07T09:02:54.4536751Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4537221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4537684Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4538069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4538446Z layer_outputs = layer( 2025-09-07T09:02:54.4538807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4539188Z attn_outputs = self.attention( 2025-09-07T09:02:54.4539604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4539995Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4540392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T09:02:54.4540839Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:02:54.4541047Z 2025-09-07T09:02:54.4541152Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4541624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4542073Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4542463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4542830Z layer_outputs = layer( 2025-09-07T09:02:54.4543204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4543581Z attn_outputs = self.attention( 2025-09-07T09:02:54.4543942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4544353Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4544741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T09:02:54.4545184Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:02:54.4545373Z 2025-09-07T09:02:54.4545449Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4545648Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4545842Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4546036Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4546255Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4546719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4547174Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4547565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4547941Z layer_outputs = layer( 2025-09-07T09:02:54.4548293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4548671Z attn_outputs = self.attention( 2025-09-07T09:02:54.4549044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4549441Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4549830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T09:02:54.4550306Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4550813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4551274Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4551462Z 2025-09-07T09:02:54.4551546Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4551756Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4551976Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4552445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4552924Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4553313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4553683Z layer_outputs = layer( 2025-09-07T09:02:54.4554034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4554428Z attn_outputs = self.attention( 2025-09-07T09:02:54.4554803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4555199Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4555583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T09:02:54.4556000Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:02:54.4556190Z 2025-09-07T09:02:54.4556292Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4556758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4557235Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4557624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4557989Z layer_outputs = layer( 2025-09-07T09:02:54.4558349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4558725Z attn_outputs = self.attention( 2025-09-07T09:02:54.4559098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4559486Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4559879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T09:02:54.4560292Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:02:54.4560450Z 2025-09-07T09:02:54.4560556Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4561026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4561479Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4561871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4562243Z layer_outputs = layer( 2025-09-07T09:02:54.4562601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4562982Z attn_outputs = self.attention( 2025-09-07T09:02:54.4563345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4563734Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4564130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T09:02:54.4564627Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T09:02:54.4565162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T09:02:54.4565639Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T09:02:54.4565833Z 2025-09-07T09:02:54.4565910Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4566131Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4566333Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4566519Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4566715Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4566912Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4567127Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4567312Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4567535Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4568006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4568498Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4568895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4569302Z layer_outputs = layer( 2025-09-07T09:02:54.4569676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4570064Z attn_outputs = self.attention( 2025-09-07T09:02:54.4570444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4570867Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4571257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:02:54.4571710Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:02:54.4572165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T09:02:54.4572633Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T09:02:54.4572831Z 2025-09-07T09:02:54.4572936Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4573397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4573859Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4574248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4574620Z layer_outputs = layer( 2025-09-07T09:02:54.4574976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4575344Z attn_outputs = self.attention( 2025-09-07T09:02:54.4575716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4576108Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4576496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:02:54.4576931Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:02:54.4577377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T09:02:54.4577824Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T09:02:54.4578012Z 2025-09-07T09:02:54.4578108Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4578576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4579030Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4579437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4579808Z layer_outputs = layer( 2025-09-07T09:02:54.4580161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4580559Z attn_outputs = self.attention( 2025-09-07T09:02:54.4580927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4581321Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4581714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:02:54.4582153Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:02:54.4582618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T09:02:54.4583023Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T09:02:54.4583172Z 2025-09-07T09:02:54.4583270Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4583739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4584225Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4584623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4585002Z layer_outputs = layer( 2025-09-07T09:02:54.4585371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4585760Z attn_outputs = self.attention( 2025-09-07T09:02:54.4586147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4586550Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4586943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:02:54.4587443Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:02:54.4588022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:02:54.4588521Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T09:02:54.4588959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T09:02:54.4589433Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T09:02:54.4589641Z 2025-09-07T09:02:54.4589721Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4589956Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4590430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4590901Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4591295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4591681Z layer_outputs = layer( 2025-09-07T09:02:54.4592056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4592443Z attn_outputs = self.attention( 2025-09-07T09:02:54.4592844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4593235Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4593626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:02:54.4594106Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:02:54.4594683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:02:54.4595167Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T09:02:54.4595583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T09:02:54.4595993Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T09:02:54.4596137Z 2025-09-07T09:02:54.4596229Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4596435Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4596657Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4597126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4597595Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4597978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4598344Z layer_outputs = layer( 2025-09-07T09:02:54.4598688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4599061Z attn_outputs = self.attention( 2025-09-07T09:02:54.4599435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4599826Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4600216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4600613Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4601014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T09:02:54.4601507Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4602021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4602477Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4602667Z 2025-09-07T09:02:54.4602746Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4602948Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4603174Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4603639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4604101Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4604489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4604868Z layer_outputs = layer( 2025-09-07T09:02:54.4605224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4605599Z attn_outputs = self.attention( 2025-09-07T09:02:54.4605966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4606373Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4606767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4607177Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4607609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T09:02:54.4608096Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4608608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:02:54.4609000Z return torch.cat(slices, dim=3) 2025-09-07T09:02:54.4609131Z 2025-09-07T09:02:54.4609208Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4609437Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4609631Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4609860Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4610330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4610813Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4611208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4611572Z layer_outputs = layer( 2025-09-07T09:02:54.4611931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4612309Z attn_outputs = self.attention( 2025-09-07T09:02:54.4612680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4613064Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4613451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4613850Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4614248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T09:02:54.4614714Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4615196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4615655Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4615846Z 2025-09-07T09:02:54.4615921Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4616145Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4616611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4617065Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4617455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4617827Z layer_outputs = layer( 2025-09-07T09:02:54.4618182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4618557Z attn_outputs = self.attention( 2025-09-07T09:02:54.4618921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4619312Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4619720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4620120Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4620519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T09:02:54.4621007Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4621498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:02:54.4621889Z return torch.cat(slices, dim=3) 2025-09-07T09:02:54.4622010Z 2025-09-07T09:02:54.4622114Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4622594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4623047Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4623445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4623835Z layer_outputs = layer( 2025-09-07T09:02:54.4624193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4624569Z attn_outputs = self.attention( 2025-09-07T09:02:54.4624931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4625323Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4625712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T09:02:54.4626203Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T09:02:54.4626733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T09:02:54.4627159Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T09:02:54.4627307Z 2025-09-07T09:02:54.4627405Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4627871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4628326Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4628720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4629084Z layer_outputs = layer( 2025-09-07T09:02:54.4629442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4629818Z attn_outputs = self.attention( 2025-09-07T09:02:54.4630188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4630574Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4630961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4631367Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4631764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T09:02:54.4632207Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:02:54.4632396Z 2025-09-07T09:02:54.4632493Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4632830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4632917Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4633179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4633257Z layer_outputs = layer( 2025-09-07T09:02:54.4633514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4633583Z attn_outputs = self.attention( 2025-09-07T09:02:54.4633834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4633927Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4634195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4634299Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4634549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T09:02:54.4634745Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:02:54.4635021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:02:54.4635159Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:02:54.4635162Z 2025-09-07T09:02:54.4635246Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4635345Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4635666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4635751Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4636000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4636072Z layer_outputs = layer( 2025-09-07T09:02:54.4636321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4636398Z attn_outputs = self.attention( 2025-09-07T09:02:54.4636646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4636736Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4636981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4637078Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4637334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T09:02:54.4637442Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:02:54.4637447Z 2025-09-07T09:02:54.4637552Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4637865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4637959Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4638206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4638271Z layer_outputs = layer( 2025-09-07T09:02:54.4638544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4638613Z attn_outputs = self.attention( 2025-09-07T09:02:54.4638869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4638969Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4639214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:02:54.4639318Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:02:54.4639566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T09:02:54.4639680Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:02:54.4639683Z 2025-09-07T09:02:54.4639760Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4639882Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4640197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4640281Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4640555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4640619Z layer_outputs = layer( 2025-09-07T09:02:54.4640875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:02:54.4640943Z attn_outputs = self.attention( 2025-09-07T09:02:54.4641195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:02:54.4641284Z self_attention_outputs = self.self_attention( 2025-09-07T09:02:54.4641533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T09:02:54.4641732Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T09:02:54.4642022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T09:02:54.4642169Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T09:02:54.4642172Z 2025-09-07T09:02:54.4642248Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4642319Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4642397Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4642466Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4642543Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4642616Z cudagraph partition due to non gpu ops 2025-09-07T09:02:54.4642714Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4643034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4643115Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4643375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:02:54.4643437Z layer_outputs = layer( 2025-09-07T09:02:54.4643683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:02:54.4643806Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:02:54.4643810Z 2025-09-07T09:02:54.4643905Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:02:54.4644251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:02:54.4644337Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:02:54.4644589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1753, in forward 2025-09-07T09:02:54.4644711Z return torch.cat([attn_output, hidden_states], dim=-1) 2025-09-07T09:02:54.4644714Z 2025-09-07T09:03:06.3614004Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:06.3614705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2495, in torch_dynamo_resume_in_forward_at_2481 2025-09-07T09:03:06.3615263Z logits = self.lm_head(sequence_output) 2025-09-07T09:03:06.3615747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1886, in forward 2025-09-07T09:03:06.3616696Z return apply_chunking_to_forward(self.forward_chunk, self.chunk_size_lm_head, self.seq_len_dim, hidden_states) 2025-09-07T09:03:06.3617341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:03:06.3617781Z return forward_fn(*input_tensors) 2025-09-07T09:03:06.3621077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1889, in forward_chunk 2025-09-07T09:03:06.3621496Z hidden_states = self.decoder(hidden_states) 2025-09-07T09:03:06.3621636Z 2025-09-07T09:03:06.3621727Z cudagraph partition due to non gpu ops 2025-09-07T09:03:06.3621934Z cudagraph partition due to non gpu ops 2025-09-07T09:03:07.2629786Z 2025-09-07T09:03:07.4854562Z running benchmark: 0% 0/30 [00:00> $GITHUB_ENV 2025-09-07T09:36:05.5719949Z echo "DEVICE_TYPE=$DEVICE_TYPE" >> $GITHUB_ENV 2025-09-07T09:36:05.5728603Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:05.5728871Z env: 2025-09-07T09:36:05.5729053Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:05.5729376Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:05.5729702Z ##[endgroup] 2025-09-07T09:36:05.5756206Z + [[ -n '' ]] 2025-09-07T09:36:05.5756525Z + python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-09-07T09:36:05.7375463Z Defaulting to user installation because normal site-packages is not writeable 2025-09-07T09:36:06.4293586Z Collecting boto3==1.35.33 2025-09-07T09:36:06.4438644Z Downloading boto3-1.35.33-py3-none-any.whl (139 kB) 2025-09-07T09:36:06.6402507Z Collecting psutil==7.0.0 2025-09-07T09:36:06.6430370Z Downloading psutil-7.0.0-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (277 kB) 2025-09-07T09:36:06.6638297Z Collecting pynvml==12.0.0 2025-09-07T09:36:06.6663439Z Downloading pynvml-12.0.0-py3-none-any.whl (26 kB) 2025-09-07T09:36:07.4220938Z Collecting botocore<1.36.0,>=1.35.33 2025-09-07T09:36:07.4248321Z Downloading botocore-1.35.99-py3-none-any.whl (13.3 MB) 2025-09-07T09:36:07.5100577Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.33) (0.10.0) 2025-09-07T09:36:07.5368727Z Collecting s3transfer<0.11.0,>=0.10.0 2025-09-07T09:36:07.5398889Z Downloading s3transfer-0.10.4-py3-none-any.whl (83 kB) 2025-09-07T09:36:07.5743025Z Collecting nvidia-ml-py<13.0.0a0,>=12.0.0 2025-09-07T09:36:07.5770019Z Downloading nvidia_ml_py-12.575.51-py3-none-any.whl (47 kB) 2025-09-07T09:36:07.5849360Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.25.10) 2025-09-07T09:36:07.5856851Z Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (2.8.1) 2025-09-07T09:36:07.7394209Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil<3.0.0,>=2.1->botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.15.0) 2025-09-07T09:36:07.8285872Z Installing collected packages: botocore, s3transfer, nvidia-ml-py, pynvml, psutil, boto3 2025-09-07T09:36:08.1494730Z Attempting uninstall: nvidia-ml-py 2025-09-07T09:36:08.1495131Z Found existing installation: nvidia-ml-py 11.525.84 2025-09-07T09:36:08.1502307Z Uninstalling nvidia-ml-py-11.525.84: 2025-09-07T09:36:08.1616418Z Successfully uninstalled nvidia-ml-py-11.525.84 2025-09-07T09:36:08.2077915Z Attempting uninstall: psutil 2025-09-07T09:36:08.2078250Z Found existing installation: psutil 5.9.8 2025-09-07T09:36:08.2118984Z Uninstalling psutil-5.9.8: 2025-09-07T09:36:08.2122931Z Successfully uninstalled psutil-5.9.8 2025-09-07T09:36:08.3355865Z Successfully installed boto3-1.35.33 botocore-1.35.99 nvidia-ml-py-12.575.51 psutil-7.0.0 pynvml-12.0.0 s3transfer-0.10.4 2025-09-07T09:36:08.4176796Z + DEVICE_NAME= 2025-09-07T09:36:08.4177075Z + DEVICE_TYPE= 2025-09-07T09:36:08.4177292Z + command -v nvidia-smi 2025-09-07T09:36:08.4177507Z + command -v rocminfo 2025-09-07T09:36:08.4177673Z + echo DEVICE_NAME= 2025-09-07T09:36:08.4177931Z + echo DEVICE_TYPE= 2025-09-07T09:36:08.4193282Z ##[group]Run set -eux 2025-09-07T09:36:08.4193470Z set -eux 2025-09-07T09:36:08.4193636Z  2025-09-07T09:36:08.4193799Z if [[ -z "${GITHUB_TOKEN}" ]]; then 2025-09-07T09:36:08.4194022Z  echo "Missing github-token input" 2025-09-07T09:36:08.4194215Z  exit 1 2025-09-07T09:36:08.4194361Z fi 2025-09-07T09:36:08.4199736Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:08.4199975Z env: 2025-09-07T09:36:08.4200127Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:08.4200478Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:08.4200760Z DEVICE_NAME: 2025-09-07T09:36:08.4200901Z DEVICE_TYPE: 2025-09-07T09:36:08.4201268Z GITHUB_TOKEN: *** 2025-09-07T09:36:08.4201423Z ##[endgroup] 2025-09-07T09:36:08.4220936Z + [[ -z *** ]] 2025-09-07T09:36:08.4248199Z ##[group]Run pytorch/test-infra/.github/actions/get-workflow-job-id@main 2025-09-07T09:36:08.4248449Z with: 2025-09-07T09:36:08.4248708Z github-token: *** 2025-09-07T09:36:08.4248866Z env: 2025-09-07T09:36:08.4249033Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:08.4249309Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:08.4249598Z DEVICE_NAME: 2025-09-07T09:36:08.4249752Z DEVICE_TYPE: 2025-09-07T09:36:08.4249900Z ##[endgroup] 2025-09-07T09:36:08.4258434Z ##[group]Run set -eux 2025-09-07T09:36:08.4258614Z set -eux 2025-09-07T09:36:08.4258765Z  2025-09-07T09:36:08.4259053Z python3 "${GITHUB_ACTION_PATH}/../../scripts/get_workflow_job_id.py" "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-09-07T09:36:08.4263415Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:08.4263642Z env: 2025-09-07T09:36:08.4263796Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:08.4264074Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:08.4264353Z DEVICE_NAME: 2025-09-07T09:36:08.4264507Z DEVICE_TYPE: 2025-09-07T09:36:08.4264780Z GITHUB_TOKEN: *** 2025-09-07T09:36:08.4265078Z ##[endgroup] 2025-09-07T09:36:08.4281877Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/get-workflow-job-id/../../scripts/get_workflow_job_id.py 17525285611 i-05ba5cf17983f662f 2025-09-07T09:36:08.9658005Z setting job-id=49775585800 2025-09-07T09:36:08.9658492Z setting job-name=inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal) 2025-09-07T09:36:08.9730037Z ##[group]Run set -eux 2025-09-07T09:36:08.9730229Z set -eux 2025-09-07T09:36:08.9730385Z  2025-09-07T09:36:08.9730536Z if [[ -n "" ]]; then 2025-09-07T09:36:08.9730719Z  source "" 2025-09-07T09:36:08.9730870Z fi 2025-09-07T09:36:08.9731019Z  2025-09-07T09:36:08.9731263Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_metadata.py" \ 2025-09-07T09:36:08.9731579Z  --schema-version "${SCHEMA_VERSION}" \ 2025-09-07T09:36:08.9731789Z  --repo "${REPO}" \ 2025-09-07T09:36:08.9731994Z  --head-branch "${HEAD_BRANCH}" \ 2025-09-07T09:36:08.9732205Z  --head-sha "${HEAD_SHA}" \ 2025-09-07T09:36:08.9732417Z  --workflow-id "${WORKFLOW_RUN_ID}" \ 2025-09-07T09:36:08.9732632Z  --run-attempt "${RUN_ATTEMPT}" \ 2025-09-07T09:36:08.9732836Z  --job-id "${JOB_ID}" \ 2025-09-07T09:36:08.9733029Z  --job-name "${JOB_NAME}" 2025-09-07T09:36:08.9741991Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:08.9742227Z env: 2025-09-07T09:36:08.9742376Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:08.9742666Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:08.9742960Z DEVICE_NAME: 2025-09-07T09:36:08.9743113Z DEVICE_TYPE: 2025-09-07T09:36:08.9743261Z SCHEMA_VERSION: v3 2025-09-07T09:36:08.9743427Z REPO: pytorch/pytorch 2025-09-07T09:36:08.9743601Z HEAD_BRANCH: refs/heads/main 2025-09-07T09:36:08.9743813Z HEAD_SHA: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T09:36:08.9744024Z WORKFLOW_RUN_ID: 17525285611 2025-09-07T09:36:08.9744197Z RUN_ATTEMPT: 1 2025-09-07T09:36:08.9744347Z JOB_ID: 49775585800 2025-09-07T09:36:08.9744677Z JOB_NAME: inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal) 2025-09-07T09:36:08.9745013Z ##[endgroup] 2025-09-07T09:36:08.9764925Z + [[ -n '' ]] 2025-09-07T09:36:08.9766164Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_metadata.py --schema-version v3 --repo pytorch/pytorch --head-branch refs/heads/main --head-sha 93fb23d6fae7c4e82c4239a1033e522088742634 --workflow-id 17525285611 --run-attempt 1 --job-id 49775585800 --job-name 'inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal)' 2025-09-07T09:36:08.9992955Z ##[group]Run set -eux 2025-09-07T09:36:08.9993140Z set -eux 2025-09-07T09:36:08.9993298Z  2025-09-07T09:36:08.9993450Z if [[ -n "" ]]; then 2025-09-07T09:36:08.9993628Z  source "" 2025-09-07T09:36:08.9993780Z fi 2025-09-07T09:36:08.9993911Z  2025-09-07T09:36:08.9994154Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_runners_info.py" 2025-09-07T09:36:08.9997673Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:08.9997906Z env: 2025-09-07T09:36:08.9998050Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:08.9998328Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:08.9998613Z DEVICE_NAME: 2025-09-07T09:36:08.9998768Z DEVICE_TYPE: 2025-09-07T09:36:08.9998910Z ##[endgroup] 2025-09-07T09:36:09.0013483Z + [[ -n '' ]] 2025-09-07T09:36:09.0013946Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_runners_info.py 2025-09-07T09:36:09.0293552Z INFO:root:Fail to import torch to get the device name 2025-09-07T09:36:09.0366967Z ##[group]Run set -eux 2025-09-07T09:36:09.0367147Z set -eux 2025-09-07T09:36:09.0367300Z  2025-09-07T09:36:09.0367465Z # TODO (huydhn): Implement this part 2025-09-07T09:36:09.0367704Z echo "dependencies={}" >> "${GITHUB_OUTPUT}" 2025-09-07T09:36:09.0371158Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:09.0371396Z env: 2025-09-07T09:36:09.0371537Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:09.0371809Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:09.0372094Z DEVICE_NAME: 2025-09-07T09:36:09.0372248Z DEVICE_TYPE: 2025-09-07T09:36:09.0372398Z ##[endgroup] 2025-09-07T09:36:09.0387048Z + echo 'dependencies={}' 2025-09-07T09:36:09.0399355Z ##[group]Run set -eux 2025-09-07T09:36:09.0399548Z set -eux 2025-09-07T09:36:09.0399700Z  2025-09-07T09:36:09.0399865Z if [[ -n "" ]]; then 2025-09-07T09:36:09.0400031Z  source "" 2025-09-07T09:36:09.0400179Z fi 2025-09-07T09:36:09.0400322Z  2025-09-07T09:36:09.0400497Z if [[ ! -d "${BENCHMARK_RESULTS_DIR}" ]]; then 2025-09-07T09:36:09.0400752Z  echo "${BENCHMARK_RESULTS_DIR} does not exist, skipping" 2025-09-07T09:36:09.0401032Z  # We don't want the job to fail if the directory doesn't exist 2025-09-07T09:36:09.0401313Z  exit 0 2025-09-07T09:36:09.0401454Z fi 2025-09-07T09:36:09.0401579Z  2025-09-07T09:36:09.0401736Z if [[ "${DRY_RUN}" == "true" ]]; then 2025-09-07T09:36:09.0402022Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-09-07T09:36:09.0402343Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-09-07T09:36:09.0402596Z  --metadata "${BENCHMARK_METADATA}" \ 2025-09-07T09:36:09.0402809Z  --runners "${RUNNER_INFO}" \ 2025-09-07T09:36:09.0403027Z  --dependencies "${DEPENDENCIES}" \ 2025-09-07T09:36:09.0403233Z  --dry-run 2025-09-07T09:36:09.0403394Z else 2025-09-07T09:36:09.0403623Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-09-07T09:36:09.0403938Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-09-07T09:36:09.0404247Z  --metadata "${BENCHMARK_METADATA}" \ 2025-09-07T09:36:09.0404462Z  --runners "${RUNNER_INFO}" \ 2025-09-07T09:36:09.0404663Z  --dependencies "${DEPENDENCIES}" 2025-09-07T09:36:09.0404857Z fi 2025-09-07T09:36:09.0408556Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:09.0408783Z env: 2025-09-07T09:36:09.0408931Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:09.0409222Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:09.0409532Z DEVICE_NAME: 2025-09-07T09:36:09.0409681Z DEVICE_TYPE: 2025-09-07T09:36:09.0409855Z BENCHMARK_RESULTS_DIR: test/test-reports 2025-09-07T09:36:09.0410056Z DRY_RUN: false 2025-09-07T09:36:09.0410840Z BENCHMARK_METADATA: {"timestamp": 1757237768, "schema_version": "v3", "name": "inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "93fb23d6fae7c4e82c4239a1033e522088742634", "workflow_id": 17525285611, "run_attempt": 1, "job_id": 49775585800} 2025-09-07T09:36:09.0411816Z RUNNER_INFO: [{"cpu_info": "x86_64", "cpu_count": 96, "avail_mem_in_gb": 188, "extra_info": {"hostname": "ip-10-0-9-178.ec2.internal"}, "name": "", "type": ""}] 2025-09-07T09:36:09.0412170Z DEPENDENCIES: {} 2025-09-07T09:36:09.0412333Z ##[endgroup] 2025-09-07T09:36:09.0428718Z + [[ -n '' ]] 2025-09-07T09:36:09.0428883Z + [[ ! -d test/test-reports ]] 2025-09-07T09:36:09.0429064Z + [[ false == \t\r\u\e ]] 2025-09-07T09:36:09.0430695Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/upload_benchmark_results.py --benchmark-results-dir test/test-reports --metadata '{"timestamp": 1757237768, "schema_version": "v3", "name": "inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "93fb23d6fae7c4e82c4239a1033e522088742634", "workflow_id": 17525285611, "run_attempt": 1, "job_id": 49775585800}' --runners '[{"cpu_info": "x86_64", "cpu_count": 96, "avail_mem_in_gb": 188, "extra_info": {"hostname": "ip-10-0-9-178.ec2.internal"}, "name": "", "type": ""}]' --dependencies '{}' 2025-09-07T09:36:09.1410404Z INFO:root:Upload test/test-reports/inductor_no_cudagraphs_torchbench_amp_inference_cpu_x86_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525285611/49775585800/inductor_no_cudagraphs_torchbench_amp_inference_cpu_x86_accuracy.json 2025-09-07T09:36:09.1629820Z INFO:botocore.credentials:Found credentials from IAM Role: gh-ci-github-action-runners-runner-role 2025-09-07T09:36:09.3403683Z INFO:root:Upload test/test-reports/inductor_dynamic_torchbench_amp_inference_cpu_x86_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525285611/49775585800/inductor_dynamic_torchbench_amp_inference_cpu_x86_accuracy.json 2025-09-07T09:36:09.4196391Z INFO:root:Upload test/test-reports/inductor_cpp_wrapper_torchbench_amp_inference_cpu_x86_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525285611/49775585800/inductor_cpp_wrapper_torchbench_amp_inference_cpu_x86_accuracy.json 2025-09-07T09:36:09.4957506Z INFO:root:Upload test/test-reports/inductor_export_torchbench_amp_inference_cpu_x86_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525285611/49775585800/inductor_export_torchbench_amp_inference_cpu_x86_accuracy.json 2025-09-07T09:36:09.5786372Z INFO:root:Upload test/test-reports/inductor_aot_inductor_torchbench_amp_inference_cpu_x86_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525285611/49775585800/inductor_aot_inductor_torchbench_amp_inference_cpu_x86_accuracy.json 2025-09-07T09:36:09.6625048Z INFO:root:Upload test/test-reports/inductor_no_cudagraphs_torchbench_amp_inference_cpu_x86_performance.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525285611/49775585800/inductor_no_cudagraphs_torchbench_amp_inference_cpu_x86_performance.json 2025-09-07T09:36:09.7569361Z INFO:root:Upload test/test-reports/inductor_no_cudagraphs_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525285611/49775585800/inductor_no_cudagraphs_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.json 2025-09-07T09:36:09.8558297Z INFO:root:Upload test/test-reports/inductor_dynamic_torchbench_amp_inference_cpu_x86_performance.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525285611/49775585800/inductor_dynamic_torchbench_amp_inference_cpu_x86_performance.json 2025-09-07T09:36:09.9554535Z INFO:root:Upload test/test-reports/inductor_dynamic_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525285611/49775585800/inductor_dynamic_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.json 2025-09-07T09:36:10.0579478Z INFO:root:Upload test/test-reports/inductor_cpp_wrapper_torchbench_amp_inference_cpu_x86_performance.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525285611/49775585800/inductor_cpp_wrapper_torchbench_amp_inference_cpu_x86_performance.json 2025-09-07T09:36:10.1837142Z INFO:root:Upload test/test-reports/inductor_cpp_wrapper_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525285611/49775585800/inductor_cpp_wrapper_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.json 2025-09-07T09:36:10.2845295Z INFO:root:Upload test/test-reports/inductor_aot_inductor_torchbench_amp_inference_cpu_x86_performance.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525285611/49775585800/inductor_aot_inductor_torchbench_amp_inference_cpu_x86_performance.json 2025-09-07T09:36:10.4151316Z INFO:root:Upload test/test-reports/inductor_aot_inductor_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525285611/49775585800/inductor_aot_inductor_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.json 2025-09-07T09:36:10.5668114Z ##[group]Run cat test/**/*_toprint.log || true 2025-09-07T09:36:10.5668453Z cat test/**/*_toprint.log || true 2025-09-07T09:36:10.5672395Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:10.5672622Z env: 2025-09-07T09:36:10.5672773Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:10.5673052Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:10.5673330Z DEVICE_NAME: 2025-09-07T09:36:10.5673481Z DEVICE_TYPE: 2025-09-07T09:36:10.5673629Z ##[endgroup] 2025-09-07T09:36:10.5733642Z cat: 'test/**/*_toprint.log': No such file or directory 2025-09-07T09:36:10.5750040Z ##[group]Run kill "$MONITOR_SCRIPT_PID" 2025-09-07T09:36:10.5750270Z kill "$MONITOR_SCRIPT_PID" 2025-09-07T09:36:10.5753875Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:10.5754108Z env: 2025-09-07T09:36:10.5754257Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:10.5754548Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:10.5754886Z DEVICE_NAME: 2025-09-07T09:36:10.5755036Z DEVICE_TYPE: 2025-09-07T09:36:10.5755195Z MONITOR_SCRIPT_PID: 57931 2025-09-07T09:36:10.5755370Z ##[endgroup] 2025-09-07T09:36:10.5836891Z Prepare all required actions 2025-09-07T09:36:10.5837269Z Getting action download info 2025-09-07T09:36:10.7347650Z Download action repository 'seemethere/upload-artifact-s3@v5' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-09-07T09:36:10.9290360Z Download action repository 'actions/upload-artifact@v4' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-09-07T09:36:11.3500329Z ##[group]Run ./.github/actions/upload-test-artifacts 2025-09-07T09:36:11.3500552Z with: 2025-09-07T09:36:11.3500821Z file-suffix: test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800 2025-09-07T09:36:11.3501125Z s3-bucket: gha-artifacts 2025-09-07T09:36:11.3501306Z env: 2025-09-07T09:36:11.3501504Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:11.3501786Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:11.3502064Z DEVICE_NAME: 2025-09-07T09:36:11.3502214Z DEVICE_TYPE: 2025-09-07T09:36:11.3502363Z ##[endgroup] 2025-09-07T09:36:11.3517030Z ##[group]Run # Remove any previous test jsons if they exist 2025-09-07T09:36:11.3517315Z # Remove any previous test jsons if they exist 2025-09-07T09:36:11.3517543Z rm -f test-jsons-*.zip 2025-09-07T09:36:11.3517810Z zip -r "test-jsons-${FILE_SUFFIX}.zip" test/test-reports -i '*.json' 2025-09-07T09:36:11.3521189Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:11.3521416Z env: 2025-09-07T09:36:11.3521558Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:11.3521860Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:11.3522149Z DEVICE_NAME: 2025-09-07T09:36:11.3522300Z DEVICE_TYPE: 2025-09-07T09:36:11.3522567Z FILE_SUFFIX: test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800 2025-09-07T09:36:11.3522854Z ##[endgroup] 2025-09-07T09:36:11.3665975Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_amp_inference_cpu_x86_accuracy.json (deflated 99%) 2025-09-07T09:36:11.3693038Z adding: test/test-reports/inductor_dynamic_torchbench_amp_inference_cpu_x86_accuracy.json (deflated 99%) 2025-09-07T09:36:11.3719887Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_amp_inference_cpu_x86_accuracy.json (deflated 99%) 2025-09-07T09:36:11.3746519Z adding: test/test-reports/inductor_export_torchbench_amp_inference_cpu_x86_accuracy.json (deflated 99%) 2025-09-07T09:36:11.3766165Z adding: test/test-reports/inductor_aot_inductor_torchbench_amp_inference_cpu_x86_accuracy.json (deflated 99%) 2025-09-07T09:36:11.3806806Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_amp_inference_cpu_x86_performance.json (deflated 99%) 2025-09-07T09:36:11.3895146Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.json (deflated 99%) 2025-09-07T09:36:11.3934861Z adding: test/test-reports/inductor_dynamic_torchbench_amp_inference_cpu_x86_performance.json (deflated 99%) 2025-09-07T09:36:11.4022503Z adding: test/test-reports/inductor_dynamic_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.json (deflated 99%) 2025-09-07T09:36:11.4061817Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_amp_inference_cpu_x86_performance.json (deflated 99%) 2025-09-07T09:36:11.4148434Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.json (deflated 99%) 2025-09-07T09:36:11.4178241Z adding: test/test-reports/inductor_aot_inductor_torchbench_amp_inference_cpu_x86_performance.json (deflated 99%) 2025-09-07T09:36:11.4220614Z adding: test/test-reports/inductor_aot_inductor_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.json (deflated 99%) 2025-09-07T09:36:11.4239086Z ##[group]Run # Remove any previous test reports if they exist 2025-09-07T09:36:11.4239375Z # Remove any previous test reports if they exist 2025-09-07T09:36:11.4239611Z rm -f test-reports-*.zip 2025-09-07T09:36:11.4239987Z zip -r "test-reports-${FILE_SUFFIX}.zip" test/test-reports -i '*.xml' -i '*.csv' 2025-09-07T09:36:11.4243414Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:11.4243649Z env: 2025-09-07T09:36:11.4243798Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:11.4244078Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:11.4244359Z DEVICE_NAME: 2025-09-07T09:36:11.4244514Z DEVICE_TYPE: 2025-09-07T09:36:11.4244775Z FILE_SUFFIX: test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800 2025-09-07T09:36:11.4245065Z ##[endgroup] 2025-09-07T09:36:11.4287686Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_amp_inference_cpu_x86_accuracy.csv (deflated 59%) 2025-09-07T09:36:11.4288252Z adding: test/test-reports/inductor_dynamic_torchbench_amp_inference_cpu_x86_accuracy.csv (deflated 58%) 2025-09-07T09:36:11.4288734Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_amp_inference_cpu_x86_accuracy.csv (deflated 59%) 2025-09-07T09:36:11.4289208Z adding: test/test-reports/inductor_export_torchbench_amp_inference_cpu_x86_accuracy.csv (deflated 62%) 2025-09-07T09:36:11.4289677Z adding: test/test-reports/inductor_aot_inductor_torchbench_amp_inference_cpu_x86_accuracy.csv (deflated 71%) 2025-09-07T09:36:11.4290173Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_amp_inference_cpu_x86_performance.csv (deflated 50%) 2025-09-07T09:36:11.4292399Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.csv (deflated 51%) 2025-09-07T09:36:11.4293052Z adding: test/test-reports/inductor_dynamic_torchbench_amp_inference_cpu_x86_performance.csv (deflated 50%) 2025-09-07T09:36:11.4295454Z adding: test/test-reports/inductor_dynamic_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.csv (deflated 50%) 2025-09-07T09:36:11.4296058Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_amp_inference_cpu_x86_performance.csv (deflated 50%) 2025-09-07T09:36:11.4298509Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.csv (deflated 51%) 2025-09-07T09:36:11.4299043Z adding: test/test-reports/inductor_aot_inductor_torchbench_amp_inference_cpu_x86_performance.csv (deflated 52%) 2025-09-07T09:36:11.4300287Z adding: test/test-reports/inductor_aot_inductor_torchbench_amp_inference_cpu_x86_performance_compilation_metrics.csv (deflated 49%) 2025-09-07T09:36:11.4314569Z ##[group]Run # Remove any previous usage logs if they exist 2025-09-07T09:36:11.4314864Z # Remove any previous usage logs if they exist 2025-09-07T09:36:11.4315101Z rm -f logs-*.zip 2025-09-07T09:36:11.4315357Z zip "logs-${FILE_SUFFIX}.zip" 'usage_log.txt' || true 2025-09-07T09:36:11.4315663Z zip -r "logs-${FILE_SUFFIX}.zip" test/test-reports -i '*.log' || true 2025-09-07T09:36:11.4319354Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:11.4319615Z env: 2025-09-07T09:36:11.4319774Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:11.4320060Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:11.4320350Z DEVICE_NAME: 2025-09-07T09:36:11.4320509Z DEVICE_TYPE: 2025-09-07T09:36:11.4320772Z FILE_SUFFIX: test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800 2025-09-07T09:36:11.4321072Z ##[endgroup] 2025-09-07T09:36:11.4398474Z adding: usage_log.txt (deflated 96%) 2025-09-07T09:36:11.4406849Z 2025-09-07T09:36:11.4407093Z zip error: Nothing to do! (logs-test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800.zip) 2025-09-07T09:36:11.4420358Z ##[group]Run # Remove any previous debugging artifacts if they exist 2025-09-07T09:36:11.4420734Z # Remove any previous debugging artifacts if they exist 2025-09-07T09:36:11.4420984Z rm -f debug-*.zip 2025-09-07T09:36:11.4421172Z if [ -d 'test/debug' ]; then 2025-09-07T09:36:11.4421475Z  zip -r "debug-${FILE_SUFFIX}.zip" test/debug 2025-09-07T09:36:11.4421686Z fi 2025-09-07T09:36:11.4425046Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:11.4425277Z env: 2025-09-07T09:36:11.4425425Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:11.4425700Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:11.4425987Z DEVICE_NAME: 2025-09-07T09:36:11.4426135Z DEVICE_TYPE: 2025-09-07T09:36:11.4426385Z FILE_SUFFIX: test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800 2025-09-07T09:36:11.4426673Z ##[endgroup] 2025-09-07T09:36:11.4476383Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-09-07T09:36:11.4476637Z with: 2025-09-07T09:36:11.4476788Z s3-bucket: gha-artifacts 2025-09-07T09:36:11.4476995Z s3-prefix: pytorch/pytorch/17525285611/1/artifact 2025-09-07T09:36:11.4477211Z retention-days: 14 2025-09-07T09:36:11.4477368Z if-no-files-found: warn 2025-09-07T09:36:11.4477551Z path: test-jsons-*.zip 2025-09-07T09:36:11.4477718Z name: artifact 2025-09-07T09:36:11.4477872Z region: us-east-1 2025-09-07T09:36:11.4478012Z env: 2025-09-07T09:36:11.4478153Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:11.4478431Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:11.4478711Z DEVICE_NAME: 2025-09-07T09:36:11.4478852Z DEVICE_TYPE: 2025-09-07T09:36:11.4478997Z ##[endgroup] 2025-09-07T09:36:11.7026479Z NOTE: s3-prefix specified, ignoring name parameter 2025-09-07T09:36:11.7026778Z With the provided path, there will be 1 file uploaded 2025-09-07T09:36:11.7027050Z Uploading to s3 prefix: pytorch/pytorch/17525285611/1/artifact 2025-09-07T09:36:11.7055157Z Starting upload of test-jsons-test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800.zip 2025-09-07T09:36:11.8383648Z Finished upload of test-jsons-test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800.zip 2025-09-07T09:36:11.8519951Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-09-07T09:36:11.8520162Z with: 2025-09-07T09:36:11.8520321Z s3-bucket: gha-artifacts 2025-09-07T09:36:11.8520533Z s3-prefix: pytorch/pytorch/17525285611/1/artifact 2025-09-07T09:36:11.8520748Z retention-days: 14 2025-09-07T09:36:11.8520903Z if-no-files-found: error 2025-09-07T09:36:11.8521082Z path: test-reports-*.zip 2025-09-07T09:36:11.8521247Z name: artifact 2025-09-07T09:36:11.8521399Z region: us-east-1 2025-09-07T09:36:11.8521541Z env: 2025-09-07T09:36:11.8521682Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:11.8521959Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:11.8522243Z DEVICE_NAME: 2025-09-07T09:36:11.8522393Z DEVICE_TYPE: 2025-09-07T09:36:11.8522541Z ##[endgroup] 2025-09-07T09:36:12.0717568Z NOTE: s3-prefix specified, ignoring name parameter 2025-09-07T09:36:12.0717864Z With the provided path, there will be 1 file uploaded 2025-09-07T09:36:12.0718171Z Uploading to s3 prefix: pytorch/pytorch/17525285611/1/artifact 2025-09-07T09:36:12.0745699Z Starting upload of test-reports-test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800.zip 2025-09-07T09:36:12.1962337Z Finished upload of test-reports-test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800.zip 2025-09-07T09:36:12.2127399Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-09-07T09:36:12.2127617Z with: 2025-09-07T09:36:12.2127770Z s3-bucket: gha-artifacts 2025-09-07T09:36:12.2127982Z s3-prefix: pytorch/pytorch/17525285611/1/artifact 2025-09-07T09:36:12.2128200Z retention-days: 14 2025-09-07T09:36:12.2128365Z if-no-files-found: ignore 2025-09-07T09:36:12.2128535Z path: logs-*.zip 2025-09-07T09:36:12.2128763Z name: artifact 2025-09-07T09:36:12.2128917Z region: us-east-1 2025-09-07T09:36:12.2129066Z env: 2025-09-07T09:36:12.2129202Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:12.2129564Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:12.2129853Z DEVICE_NAME: 2025-09-07T09:36:12.2130005Z DEVICE_TYPE: 2025-09-07T09:36:12.2130141Z ##[endgroup] 2025-09-07T09:36:12.4370449Z NOTE: s3-prefix specified, ignoring name parameter 2025-09-07T09:36:12.4370762Z With the provided path, there will be 1 file uploaded 2025-09-07T09:36:12.4371030Z Uploading to s3 prefix: pytorch/pytorch/17525285611/1/artifact 2025-09-07T09:36:12.4398264Z Starting upload of logs-test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800.zip 2025-09-07T09:36:12.5753487Z Finished upload of logs-test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800.zip 2025-09-07T09:36:12.5888769Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-09-07T09:36:12.5889053Z with: 2025-09-07T09:36:12.5889220Z s3-bucket: gha-artifacts 2025-09-07T09:36:12.5889436Z s3-prefix: pytorch/pytorch/17525285611/1/artifact 2025-09-07T09:36:12.5889657Z retention-days: 14 2025-09-07T09:36:12.5889833Z if-no-files-found: ignore 2025-09-07T09:36:12.5890017Z path: debug-*.zip 2025-09-07T09:36:12.5890180Z name: artifact 2025-09-07T09:36:12.5890338Z region: us-east-1 2025-09-07T09:36:12.5890484Z env: 2025-09-07T09:36:12.5890637Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:12.5890921Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:12.5891211Z DEVICE_NAME: 2025-09-07T09:36:12.5891360Z DEVICE_TYPE: 2025-09-07T09:36:12.5891514Z ##[endgroup] 2025-09-07T09:36:12.8016346Z No files were found with the provided path: debug-*.zip. No artifacts will be uploaded. 2025-09-07T09:36:12.8157086Z ##[group]Run # shellcheck disable=SC2156 2025-09-07T09:36:12.8157342Z # shellcheck disable=SC2156 2025-09-07T09:36:12.8157699Z find . -iname "core.[1-9]*" -exec docker exec "${DOCKER_CONTAINER_ID}" sh -c "gdb python {} -ex 'bt' -ex 'q'" \; 2025-09-07T09:36:12.8162289Z shell: /usr/bin/bash -e {0} 2025-09-07T09:36:12.8162469Z env: 2025-09-07T09:36:12.8162619Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:12.8162894Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:12.8163186Z DEVICE_NAME: 2025-09-07T09:36:12.8163345Z DEVICE_TYPE: 2025-09-07T09:36:12.8163492Z ##[endgroup] 2025-09-07T09:36:12.9844968Z Prepare all required actions 2025-09-07T09:36:12.9845244Z Getting action download info 2025-09-07T09:36:13.0752832Z ##[group]Run ./.github/actions/upload-utilization-stats 2025-09-07T09:36:13.0753055Z with: 2025-09-07T09:36:13.0753206Z job_id: 49775585800 2025-09-07T09:36:13.0753538Z job_name: inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal) 2025-09-07T09:36:13.0753906Z workflow_name: inductor-perf-nightly-x86 2025-09-07T09:36:13.0754122Z workflow_run_id: 17525285611 2025-09-07T09:36:13.0754304Z workflow_attempt: 1 2025-09-07T09:36:13.0754460Z env: 2025-09-07T09:36:13.0754604Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:13.0754871Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:13.0755153Z DEVICE_NAME: 2025-09-07T09:36:13.0755301Z DEVICE_TYPE: 2025-09-07T09:36:13.0755446Z ##[endgroup] 2025-09-07T09:36:13.0765516Z ##[group]Run echo "workflow_id: 17525285611" 2025-09-07T09:36:13.0765751Z echo "workflow_id: 17525285611" 2025-09-07T09:36:13.0765955Z echo "workflow_attempt: 1" 2025-09-07T09:36:13.0766186Z echo "workflow_Name: inductor-perf-nightly-x86" 2025-09-07T09:36:13.0766419Z echo "job_id: 49775585800" 2025-09-07T09:36:13.0766785Z echo "job_name: inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal)" 2025-09-07T09:36:13.0767153Z echo "artifact_prefix: " 2025-09-07T09:36:13.0767410Z python3 --version 2025-09-07T09:36:13.0771402Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:13.0771625Z env: 2025-09-07T09:36:13.0771775Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:13.0772045Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:13.0772318Z DEVICE_NAME: 2025-09-07T09:36:13.0772467Z DEVICE_TYPE: 2025-09-07T09:36:13.0772617Z ##[endgroup] 2025-09-07T09:36:13.0789530Z workflow_id: 17525285611 2025-09-07T09:36:13.0789767Z workflow_attempt: 1 2025-09-07T09:36:13.0789972Z workflow_Name: inductor-perf-nightly-x86 2025-09-07T09:36:13.0790186Z job_id: 49775585800 2025-09-07T09:36:13.0790514Z job_name: inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal) 2025-09-07T09:36:13.0790867Z artifact_prefix: 2025-09-07T09:36:13.0800162Z Python 3.9.23 2025-09-07T09:36:13.0822488Z ##[group]Run nick-fields/retry@v3.0.0 2025-09-07T09:36:13.0822761Z with: 2025-09-07T09:36:13.0822901Z shell: bash 2025-09-07T09:36:13.0823049Z timeout_minutes: 5 2025-09-07T09:36:13.0823207Z max_attempts: 5 2025-09-07T09:36:13.0823367Z retry_wait_seconds: 30 2025-09-07T09:36:13.0823682Z command: set -eu python3 -m pip install python-dateutil==2.8.2 boto3==1.35.42 pandas==2.1.3 dataclasses_json==0.6.7 2025-09-07T09:36:13.0824025Z polling_interval_seconds: 1 2025-09-07T09:36:13.0824204Z warning_on_retry: true 2025-09-07T09:36:13.0824381Z continue_on_error: false 2025-09-07T09:36:13.0824539Z env: 2025-09-07T09:36:13.0824681Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:13.0824954Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:13.0825239Z DEVICE_NAME: 2025-09-07T09:36:13.0825388Z DEVICE_TYPE: 2025-09-07T09:36:13.0825524Z ##[endgroup] 2025-09-07T09:36:13.3129861Z Defaulting to user installation because normal site-packages is not writeable 2025-09-07T09:36:13.3661441Z Collecting python-dateutil==2.8.2 2025-09-07T09:36:13.3956912Z Downloading python_dateutil-2.8.2-py2.py3-none-any.whl (247 kB) 2025-09-07T09:36:14.0449547Z Collecting boto3==1.35.42 2025-09-07T09:36:14.0544635Z Downloading boto3-1.35.42-py3-none-any.whl (139 kB) 2025-09-07T09:36:14.3928836Z Collecting pandas==2.1.3 2025-09-07T09:36:14.4038331Z Downloading pandas-2.1.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (12.3 MB) 2025-09-07T09:36:14.4968509Z Requirement already satisfied: dataclasses_json==0.6.7 in /home/ec2-user/.local/lib/python3.9/site-packages (0.6.7) 2025-09-07T09:36:14.4978182Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil==2.8.2) (1.15.0) 2025-09-07T09:36:14.5009822Z Requirement already satisfied: s3transfer<0.11.0,>=0.10.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from boto3==1.35.42) (0.10.4) 2025-09-07T09:36:14.5013898Z Requirement already satisfied: botocore<1.36.0,>=1.35.42 in /home/ec2-user/.local/lib/python3.9/site-packages (from boto3==1.35.42) (1.35.99) 2025-09-07T09:36:14.5017316Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.42) (0.10.0) 2025-09-07T09:36:14.5620002Z Collecting tzdata>=2022.1 2025-09-07T09:36:14.5707817Z Downloading tzdata-2025.2-py2.py3-none-any.whl (347 kB) 2025-09-07T09:36:14.5776531Z Requirement already satisfied: pytz>=2020.1 in /usr/lib/python3.9/site-packages (from pandas==2.1.3) (2022.7.1) 2025-09-07T09:36:15.1028523Z Collecting numpy<2,>=1.22.4 2025-09-07T09:36:15.1113453Z Downloading numpy-1.26.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (18.2 MB) 2025-09-07T09:36:15.2363711Z Requirement already satisfied: marshmallow<4.0.0,>=3.18.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from dataclasses_json==0.6.7) (3.26.1) 2025-09-07T09:36:15.2364567Z Requirement already satisfied: typing-inspect<1,>=0.4.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from dataclasses_json==0.6.7) (0.9.0) 2025-09-07T09:36:15.2406458Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.42->boto3==1.35.42) (1.25.10) 2025-09-07T09:36:15.2480723Z Requirement already satisfied: packaging>=17.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from marshmallow<4.0.0,>=3.18.0->dataclasses_json==0.6.7) (25.0) 2025-09-07T09:36:15.2549423Z Requirement already satisfied: typing-extensions>=3.7.4 in /home/ec2-user/.local/lib/python3.9/site-packages (from typing-inspect<1,>=0.4.0->dataclasses_json==0.6.7) (4.15.0) 2025-09-07T09:36:15.2553940Z Requirement already satisfied: mypy-extensions>=0.3.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from typing-inspect<1,>=0.4.0->dataclasses_json==0.6.7) (1.1.0) 2025-09-07T09:36:15.3862892Z Installing collected packages: python-dateutil, tzdata, numpy, pandas, boto3 2025-09-07T09:36:18.8510665Z Attempting uninstall: boto3 2025-09-07T09:36:18.8510978Z Found existing installation: boto3 1.35.33 2025-09-07T09:36:18.8568459Z Uninstalling boto3-1.35.33: 2025-09-07T09:36:18.8576556Z Successfully uninstalled boto3-1.35.33 2025-09-07T09:36:18.8940768Z Successfully installed boto3-1.35.42 numpy-1.26.4 pandas-2.1.3 python-dateutil-2.8.2 tzdata-2025.2 2025-09-07T09:36:19.1414498Z Command completed after 1 attempt(s). 2025-09-07T09:36:19.1454470Z ##[group]Run python3 -m tools.stats.upload_utilization_stats.upload_utilization_stats \ 2025-09-07T09:36:19.1454905Z python3 -m tools.stats.upload_utilization_stats.upload_utilization_stats \ 2025-09-07T09:36:19.1455201Z  --workflow-run-id "17525285611" \ 2025-09-07T09:36:19.1455454Z  --workflow-name "inductor-perf-nightly-x86" \ 2025-09-07T09:36:19.1455723Z  --workflow-run-attempt "1" \ 2025-09-07T09:36:19.1455918Z  --job-id "49775585800" \ 2025-09-07T09:36:19.1456286Z  --job-name "inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal)" \ 2025-09-07T09:36:19.1456661Z  --local-path "" \ 2025-09-07T09:36:19.1456849Z  --artifact-prefix "" 2025-09-07T09:36:19.1461218Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:19.1461447Z env: 2025-09-07T09:36:19.1461603Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:19.1461880Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:19.1462165Z DEVICE_NAME: 2025-09-07T09:36:19.1462311Z DEVICE_TYPE: 2025-09-07T09:36:19.1462581Z ##[endgroup] 2025-09-07T09:36:20.0658363Z repo: pytorch/pytorch 2025-09-07T09:36:20.0658631Z Search for test log in s3 bucket: ossci-utilization 2025-09-07T09:36:20.0658999Z Downloading logs-test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800.zip 2025-09-07T09:36:20.0659492Z extracting usage_log.txt from zip file logs-test-inductor_torchbench_perf_cpu_x86-2-4-linux.24xl.spr-metal_49775585800.zip 2025-09-07T09:36:20.0659871Z Converted Log Model: UtilizationMetadata: 2025-09-07T09:36:20.0660742Z UtilizationMetadata(level='metadata', workflow_id='17525285611', job_id='49775585800', workflow_name='inductor-perf-nightly-x86', job_name='inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal)', usage_collect_interval=4.0, data_model_version=1.5, start_at=1757231027, gpu_count=0, cpu_count=96, gpu_type=None, error=None) 2025-09-07T09:36:20.0661611Z [Db Segments] detected pytest cmd: 15, generated segments: 15 2025-09-07T09:36:20.0661861Z [db model] Peek db timeseries 2025-09-07T09:36:20.0662041Z :{ 2025-09-07T09:36:20.0662189Z "created_at": 1757237779, 2025-09-07T09:36:20.0662371Z "type": "utilization", 2025-09-07T09:36:20.0662534Z "tags": [ 2025-09-07T09:36:20.0662682Z "record" 2025-09-07T09:36:20.0662828Z ], 2025-09-07T09:36:20.0662977Z "time_stamp": 1757231027, 2025-09-07T09:36:20.0663151Z "repo": "pytorch/pytorch", 2025-09-07T09:36:20.0663333Z "workflow_id": 17525285611, 2025-09-07T09:36:20.0663712Z "run_attempt": 1, 2025-09-07T09:36:20.0663878Z "job_id": 49775585800, 2025-09-07T09:36:20.0664075Z "workflow_name": "inductor-perf-nightly-x86", 2025-09-07T09:36:20.0664459Z "job_name": "inductor-test-nightly-freezing / test (inductor_torchbench_perf_cpu_x86, 2, 4, linux.24xl.spr-metal)", 2025-09-07T09:36:20.0664798Z "json_data": "{}" 2025-09-07T09:36:20.0664966Z } 2025-09-07T09:36:20.0665265Z Writing 1 documents to S3 ossci-utilization/util_metadata/v_1.5/pytorch/pytorch/17525285611/1/49775585800/metadata 2025-09-07T09:36:20.0665761Z Done! Finish writing document to S3 ossci-utilization/util_metadata/v_1.5/pytorch/pytorch/17525285611/1/49775585800/metadata 2025-09-07T09:36:20.0666275Z Writing 447 documents to S3 ossci-utilization/util_timeseries/v_1.5/pytorch/pytorch/17525285611/1/49775585800/time_series 2025-09-07T09:36:20.0666797Z Done! Finish writing document to S3 ossci-utilization/util_timeseries/v_1.5/pytorch/pytorch/17525285611/1/49775585800/time_series 2025-09-07T09:36:20.1444233Z ##[group]Run pytorch/test-infra/.github/actions/teardown-linux@main 2025-09-07T09:36:20.1444584Z with: 2025-09-07T09:36:20.1444777Z env: 2025-09-07T09:36:20.1444972Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:20.1445366Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:20.1445759Z DEVICE_NAME: 2025-09-07T09:36:20.1445959Z DEVICE_TYPE: 2025-09-07T09:36:20.1446146Z ##[endgroup] 2025-09-07T09:36:20.1460774Z ##[group]Run set -eou pipefail 2025-09-07T09:36:20.1461130Z set -eou pipefail 2025-09-07T09:36:20.1461365Z  2025-09-07T09:36:20.1461675Z echo "Holding runner for 2 hours until all ssh sessions have logged out" 2025-09-07T09:36:20.1462044Z for _ in $(seq 1440); do 2025-09-07T09:36:20.1462334Z  # Break if no ssh session exists anymore 2025-09-07T09:36:20.1462627Z  if [ "$(who)" = "" ]; then 2025-09-07T09:36:20.1462880Z  break 2025-09-07T09:36:20.1463133Z  fi 2025-09-07T09:36:20.1463339Z  echo "." 2025-09-07T09:36:20.1463545Z  sleep 5 2025-09-07T09:36:20.1463746Z done 2025-09-07T09:36:20.1468947Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:20.1469255Z env: 2025-09-07T09:36:20.1469451Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:20.1469831Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:20.1470213Z DEVICE_NAME: 2025-09-07T09:36:20.1470412Z DEVICE_TYPE: 2025-09-07T09:36:20.1470605Z ##[endgroup] 2025-09-07T09:36:20.1488993Z Holding runner for 2 hours until all ssh sessions have logged out 2025-09-07T09:36:20.1549174Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-09-07T09:36:20.1549525Z # ignore expansion of "docker ps -q" since it could be empty 2025-09-07T09:36:20.1549790Z # shellcheck disable=SC2046 2025-09-07T09:36:20.1550020Z docker stop $(docker ps -q) || true 2025-09-07T09:36:20.1550241Z # Prune all of the docker images 2025-09-07T09:36:20.1550454Z docker system prune -af 2025-09-07T09:36:20.1554307Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:20.1554550Z env: 2025-09-07T09:36:20.1554709Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:36:20.1555007Z DOCKER_CONTAINER_ID: 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:20.1555308Z DEVICE_NAME: 2025-09-07T09:36:20.1555474Z DEVICE_TYPE: 2025-09-07T09:36:20.1555629Z ##[endgroup] 2025-09-07T09:36:31.1259010Z 5b1006e7d104 2025-09-07T09:36:31.9597886Z Deleted Containers: 2025-09-07T09:36:31.9598959Z 5b1006e7d10405a4e64728a1885d0ff6c713a58d8f0dc2a9ea0c587915151523 2025-09-07T09:36:31.9599159Z 2025-09-07T09:36:38.3107456Z Deleted Images: 2025-09-07T09:36:38.3108130Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T09:36:38.3109133Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image@sha256:383efb45082f20b8c808cb0ba4df693a01359592233f641f1f486911ac320a9a 2025-09-07T09:36:38.3109631Z deleted: sha256:662d8c9dfc7db2f5d004293de4f2b7647941dee4c916479ef082d17fcdfd9c47 2025-09-07T09:36:38.3109991Z deleted: sha256:ea5ad443c754124b3a5a209c2663376b4c156947edef1b982a336148bbf9114d 2025-09-07T09:36:38.3110351Z deleted: sha256:284be7504f072e0c04da4e2190e8d0e1de73835ed67be81f3ddd7eafd5d06a3a 2025-09-07T09:36:38.3110708Z deleted: sha256:2f49ff4be65f7ca55de8d7028fb3df7d08232a9f043aa7ba27d9393724286281 2025-09-07T09:36:38.3111063Z deleted: sha256:f63b503fdd1cca198aecefb9eef7ffbeb5fbc723f2a8462f50316e56cd403cbc 2025-09-07T09:36:38.3111420Z deleted: sha256:f9d46e08457013f0e71d608ac3dd95b79c41120060a80baefa684048cc15574e 2025-09-07T09:36:38.3111770Z deleted: sha256:cab76e28615751b6d6a703103b1da790a67cb3a4ee2e8814de51de18ff8b595d 2025-09-07T09:36:38.3112263Z deleted: sha256:0b2d09aa482371591a32563a5db71472822abd096a347967a9bd2a177737109f 2025-09-07T09:36:38.3112732Z deleted: sha256:d306d346d5da05e9fd04284304b1637a0bf01ee97397c688d19d783d5e133de9 2025-09-07T09:36:38.3113091Z deleted: sha256:bb3381a916d410a6e304540bb0796099dc780cd11f5829e734b337e0e79acfe4 2025-09-07T09:36:38.3113428Z deleted: sha256:bcf487c27e826c092985285163fb896e3324460b1774f3eb2a66623cd31e7d87 2025-09-07T09:36:38.3113781Z deleted: sha256:7d13485a9bdc5c0e64ac5085b25f4dded75c60f74090369c1b6f3f546ee37e94 2025-09-07T09:36:38.3114134Z deleted: sha256:55351d98a4197542fa7c78089671f447a6ef88cc554b7fad4fc522e8d4d187b6 2025-09-07T09:36:38.3114483Z deleted: sha256:f884bc0c4f9a994f3b3f1d82205f3a7014b05c84ad0c1c2fa3254d15a44f31e1 2025-09-07T09:36:38.3114828Z deleted: sha256:cdd16785a15239e518604ea9ea31405d5225fa6411d1c6d74d6523bcebf759ab 2025-09-07T09:36:38.3115183Z deleted: sha256:2c5bc1dc49446d7df5784578ae7c99460a93b502aa0c3b9deffbb95ec5216860 2025-09-07T09:36:38.3115546Z deleted: sha256:bae1e956be98416ce7d1a6c2c6ef0917f467238e19291786f8e1fed36fa81956 2025-09-07T09:36:38.3115906Z deleted: sha256:2cb1f002ab1126b0606999a9557b3f7f5da1e453d5376d29d95d60a979a215c4 2025-09-07T09:36:38.3116259Z deleted: sha256:25055a5f67b9bce8fac50ee1508dcb0f862ed154de5ded734e55f60edaca385f 2025-09-07T09:36:38.3116611Z deleted: sha256:98024e2dd34a5899240e41ae14f59c657cdc005040773e6ad7cfe3d67cdac7a8 2025-09-07T09:36:38.3116966Z deleted: sha256:8d2e75659096b4af8a20c3e9a6cce899b6e720f638eacdfd7d41ec8a736efdde 2025-09-07T09:36:38.3117313Z deleted: sha256:7741a6bf043548509c51c32e44734f30dfe07f91ca56c64422b004c3c0444e68 2025-09-07T09:36:38.3117665Z deleted: sha256:e2e63edbd2512e413c388888eabade05a2a7876adf20e7f0e0c3660ac3acbd3d 2025-09-07T09:36:38.3118024Z deleted: sha256:7fdea0f7711ee22084f87dc6d651598b5e5c5237de828105f698cb6a937d4c9c 2025-09-07T09:36:38.3118373Z deleted: sha256:486a2cf42f9492f291d59d48f3cec5a0a72449d8b6ad7d7a02596da237cdd154 2025-09-07T09:36:38.3118728Z deleted: sha256:a17da64c93a4939fad81a3ff6b6cb30f988176a6e0062fcf9c65e06cd9b9c3fb 2025-09-07T09:36:38.3119097Z deleted: sha256:70b4a3a917b8f95b19ae5dab6f404af8fa1c886022e4a1d785654013d5d876af 2025-09-07T09:36:38.3119457Z deleted: sha256:bd1b9d6a8aa636a67023800dcd85e4a3a7a7a21d65c6e6491d169fa65b4404a9 2025-09-07T09:36:38.3119806Z deleted: sha256:e3befcf3d3693c1d7bf0535e6e6722f0aabb0123805443ef5915dd5441ed0b00 2025-09-07T09:36:38.3120161Z deleted: sha256:4b4f846f1c4266b015f5fdf8dac5346c083c3aee2375e337172c112677c5a8c0 2025-09-07T09:36:38.3120513Z deleted: sha256:f05dc4d1350267b90e07af241a64f86a928fb3d8de75717ac04ec5a0433d042f 2025-09-07T09:36:38.3120869Z deleted: sha256:b6b4de696915fa2db09844ec9ac44dbb2940b655cd356404cf1ff03eec644dad 2025-09-07T09:36:38.3121223Z deleted: sha256:da008bbe1fc29cb35b3949040e97eb801f3264a56c4dd1b9d43a3cb54f2a39b2 2025-09-07T09:36:38.3121578Z deleted: sha256:261da5d14cad99ee11dcdaeb6055726f38fc12b7c559ee9c6d2ddc3f288f4828 2025-09-07T09:36:38.3121938Z deleted: sha256:16f900c60e70d685a85ca571ee0dada993a02217bdd6bb8b1d49169e7e28cf41 2025-09-07T09:36:38.3122297Z deleted: sha256:f57b18c5cde1d1dc553a15e1e98141d4afc0b4d0bb1182cc85b2c21bd18bb783 2025-09-07T09:36:38.3122688Z deleted: sha256:3c79105088ac60b231e4553752ee42cb6a87f9d32736b32f0c2123dddec724e7 2025-09-07T09:36:38.3123031Z deleted: sha256:df1ffff478908236efb6ceb8e05e6e078f12b864f4d24ce598cba7b961fad65c 2025-09-07T09:36:38.3123385Z deleted: sha256:8170255b562b59b76768f18a5b84b1ba887db93d3fe43b87a74bdc6be4f82014 2025-09-07T09:36:38.3123736Z deleted: sha256:c863cfe6bed704be5a54617331e27158b6f5a492dd6b9ed9c99d23db017cf5e1 2025-09-07T09:36:38.3124091Z deleted: sha256:e9e5a98c073f72c3abf9cc98724a31a3791535574ac78aeda7eb5df4580b21d0 2025-09-07T09:36:38.3124441Z deleted: sha256:0a42ac98735ca6578911218be7a7918001fe8aee1eb33d98f0d0a153d0e1102d 2025-09-07T09:36:38.3124785Z deleted: sha256:77d5a8aaa4d0fe1210dda9ac1f0fa3cf6141fea925b6240b9839d7505d021d3f 2025-09-07T09:36:38.3125138Z deleted: sha256:fa6ec46c43532dc01449df1cc403de8bb5872f859076e90658534c51c1487ef9 2025-09-07T09:36:38.3125568Z deleted: sha256:424a12dd5083283e19af48d31b7f2e33911ca8f459796f17280eaf5777a9aa25 2025-09-07T09:36:38.3125940Z deleted: sha256:8f0499601e14f1073e20ce889b45d12ab33264f9cf30359ac29dddbf58a311aa 2025-09-07T09:36:38.3126293Z deleted: sha256:5a5fae32dfb81abcd7bf374018b11e8e42a5aa39841d4b94e822d306c9af015b 2025-09-07T09:36:38.3126643Z deleted: sha256:d1bda89f22d383d38dfb7f7590b3bb202ccb91814034e7c7e2493306a10151ef 2025-09-07T09:36:38.3127003Z deleted: sha256:dbf16c1fcae146528685a8f745f9c505b24ba9ef009c42b1bd711ff7bf51b936 2025-09-07T09:36:38.3127357Z deleted: sha256:f9ec0065788f638325536a37427e2635b760a32457f20ca0acbcef6946b1041b 2025-09-07T09:36:38.3127704Z deleted: sha256:9d9911dac8fb2ff7db87329f38625d73f452dfef8822830048bbc00541c7df14 2025-09-07T09:36:38.3128050Z deleted: sha256:de4c1937129850e357b0de484d230569f628ac0bc883b12eff42932cd1e193ce 2025-09-07T09:36:38.3128400Z deleted: sha256:7b3c9e5b56a1d74226a5c1a54e5cb5e749012aa9b1d2376c6e7503757e29c35b 2025-09-07T09:36:38.3128754Z deleted: sha256:8062a6f28fc5fe2a199e1c1c40b6c43b7e29eb0c452492b47ec6900413b19cb6 2025-09-07T09:36:38.3129116Z deleted: sha256:f879aeffe6886f8da80462b571f9307aa63bb961645bec55ff579187a81cfd0b 2025-09-07T09:36:38.3129484Z deleted: sha256:5c6ef06b3536a430194aee509a784ee889c4a9d6248cb20fd9290e87e4ee2245 2025-09-07T09:36:38.3129837Z deleted: sha256:461aea034a25a2d72be6adfe9213c457c4cbf48724e9cb1c57987afb87668f21 2025-09-07T09:36:38.3130204Z deleted: sha256:e342cd1c71b7d0b024ea16b4a11f3f7fbbc2e3d11ef754c9d242aa50c4f8b0a3 2025-09-07T09:36:38.3130581Z deleted: sha256:bffd35a7fa1ddcfe05f79b7d3cae4180928eeea00eaab7ed7f484bc31adfc1d5 2025-09-07T09:36:38.3130952Z deleted: sha256:b34e33e7b04b5cbb5d5852199430593bfa18ddfe9081df42284230a14ebb739e 2025-09-07T09:36:38.3131311Z deleted: sha256:21d9b55338774d9ddc66d0bfcc92af9c8d2ecd94d1710b7049f5a811e411af7b 2025-09-07T09:36:38.3131662Z deleted: sha256:6cc2b33909585d17bf269fb8297ff881249e136137254734f7d23b9583208718 2025-09-07T09:36:38.3132020Z deleted: sha256:ca7f55b7c6d6cb11ddd8e187da34c2695fc2ce7655d652b9c9dc140a01ed056f 2025-09-07T09:36:38.3132394Z deleted: sha256:a3ece3d0ab6e99ef783c4f8d27d0e38504ab4477590ef556c16d22d92ba63a43 2025-09-07T09:36:38.3132743Z deleted: sha256:c137b0d41177c753aa1b69b11d0dd1f82420bf8520371866c845b53dca10b2d0 2025-09-07T09:36:38.3133082Z deleted: sha256:1e0d92b07bce12e511af59f608edd1932b10704d700f5e7538e406b90ecbb615 2025-09-07T09:36:38.3133436Z deleted: sha256:2ec3d01b3031e9da124d67410f54866ec5c679a0d6e4aee6b31608c45ce7fd77 2025-09-07T09:36:38.3133780Z deleted: sha256:308cffbd71363688c672b2043c6b9bf647cfb84593c42c3d88e3f36ee8f7f1b4 2025-09-07T09:36:38.3134125Z deleted: sha256:d965d9873fa450daba50a85d961f0835b14374167d84cfafa6060d16229f4229 2025-09-07T09:36:38.3134478Z deleted: sha256:effd997e222f62a34133bb2ecf9c0ffee151e5797f72e734d86a270d2e722374 2025-09-07T09:36:38.3134833Z deleted: sha256:0bbc1c78c10ee09c2697cfcce347dc9edbf82a7ccc25a6db6ee0a8dda398f7f2 2025-09-07T09:36:38.3135191Z deleted: sha256:214858e773d1ad73c2965c19b29cbfd3e2a974daa879163e1c1eb96567a7ee06 2025-09-07T09:36:38.3135545Z deleted: sha256:a9c7a2cd7ae229b26e84c093de657d0f4334d6cc9301991c6c3245ff62a9a71d 2025-09-07T09:36:38.3135919Z deleted: sha256:749a80551ef3f272e2517cb065bc7a5250da47d0b36bf74ed453caa9a5fee265 2025-09-07T09:36:38.3136260Z deleted: sha256:39b014c4e62d21c11df6c6d775d3f345675014292198981f455bacc4515a0f7b 2025-09-07T09:36:38.3136607Z deleted: sha256:0f087c9a894566644f825f5f87308d92e4cf149c51f7cd4769cbfaeefd3df791 2025-09-07T09:36:38.3136958Z deleted: sha256:dc6eb6dad5f9e332f00af553440e857b1467db1be43dd910cdb6830ba0898d50 2025-09-07T09:36:38.3137165Z 2025-09-07T09:36:38.3137282Z Total reclaimed space: 68.72GB 2025-09-07T09:36:38.3206541Z Post job cleanup. 2025-09-07T09:36:38.3236405Z Post job cleanup. 2025-09-07T09:36:38.3941935Z [command]/usr/bin/git version 2025-09-07T09:36:38.3972071Z git version 2.47.1 2025-09-07T09:36:38.3996171Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/ae1c423d-399d-47ff-a641-a8b803fa7d9c/.gitconfig' 2025-09-07T09:36:38.4002202Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/ae1c423d-399d-47ff-a641-a8b803fa7d9c' before making global git config changes 2025-09-07T09:36:38.4002926Z Adding repository directory to the temporary git global config as a safe directory 2025-09-07T09:36:38.4005702Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T09:36:38.4039525Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-09-07T09:36:38.4067859Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-09-07T09:36:38.4318682Z Entering 'android/libs/fbjni' 2025-09-07T09:36:38.4358665Z Entering 'third_party/FP16' 2025-09-07T09:36:38.4403422Z Entering 'third_party/FXdiv' 2025-09-07T09:36:38.4446310Z Entering 'third_party/NNPACK' 2025-09-07T09:36:38.4489195Z Entering 'third_party/NVTX' 2025-09-07T09:36:38.4535555Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T09:36:38.4582484Z Entering 'third_party/XNNPACK' 2025-09-07T09:36:38.4643262Z Entering 'third_party/aiter' 2025-09-07T09:36:38.4687731Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T09:36:38.4732654Z Entering 'third_party/benchmark' 2025-09-07T09:36:38.4776305Z Entering 'third_party/composable_kernel' 2025-09-07T09:36:38.4827384Z Entering 'third_party/cpp-httplib' 2025-09-07T09:36:38.4872373Z Entering 'third_party/cpuinfo' 2025-09-07T09:36:38.4914455Z Entering 'third_party/cudnn_frontend' 2025-09-07T09:36:38.4956235Z Entering 'third_party/cutlass' 2025-09-07T09:36:38.5006450Z Entering 'third_party/fbgemm' 2025-09-07T09:36:38.5053265Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T09:36:38.5099205Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T09:36:38.5148019Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T09:36:38.5189236Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T09:36:38.5236284Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T09:36:38.5280123Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T09:36:38.5319071Z Entering 'third_party/fbgemm/external/json' 2025-09-07T09:36:38.5365579Z Entering 'third_party/flash-attention' 2025-09-07T09:36:38.5409817Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T09:36:38.5456901Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T09:36:38.5504088Z Entering 'third_party/flatbuffers' 2025-09-07T09:36:38.5549702Z Entering 'third_party/fmt' 2025-09-07T09:36:38.5594813Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T09:36:38.5636800Z Entering 'third_party/gloo' 2025-09-07T09:36:38.5683649Z Entering 'third_party/googletest' 2025-09-07T09:36:38.5726639Z Entering 'third_party/ideep' 2025-09-07T09:36:38.5771097Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T09:36:38.5817123Z Entering 'third_party/ittapi' 2025-09-07T09:36:38.5857800Z Entering 'third_party/kineto' 2025-09-07T09:36:38.5899277Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T09:36:38.5939431Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T09:36:38.5981833Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T09:36:38.6024897Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T09:36:38.6072291Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T09:36:38.6113808Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T09:36:38.6161727Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T09:36:38.6206941Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T09:36:38.6250969Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T09:36:38.6291987Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T09:36:38.6335249Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T09:36:38.6378913Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T09:36:38.6422637Z Entering 'third_party/kleidiai' 2025-09-07T09:36:38.6462484Z Entering 'third_party/mimalloc' 2025-09-07T09:36:38.6507516Z Entering 'third_party/nlohmann' 2025-09-07T09:36:38.6553966Z Entering 'third_party/onnx' 2025-09-07T09:36:38.6608521Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T09:36:38.6650049Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T09:36:38.6693252Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T09:36:38.6733764Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T09:36:38.6777587Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T09:36:38.6819515Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T09:36:38.6864330Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T09:36:38.6909708Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T09:36:38.6953381Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T09:36:38.6997334Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T09:36:38.7039394Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T09:36:38.7082730Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T09:36:38.7142375Z Entering 'third_party/pocketfft' 2025-09-07T09:36:38.7183994Z Entering 'third_party/protobuf' 2025-09-07T09:36:38.7230523Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T09:36:38.7274753Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T09:36:38.7318982Z Entering 'third_party/psimd' 2025-09-07T09:36:38.7364104Z Entering 'third_party/pthreadpool' 2025-09-07T09:36:38.7407998Z Entering 'third_party/pybind11' 2025-09-07T09:36:38.7453587Z Entering 'third_party/python-peachpy' 2025-09-07T09:36:38.7499410Z Entering 'third_party/sleef' 2025-09-07T09:36:38.7544986Z Entering 'third_party/tensorpipe' 2025-09-07T09:36:38.7588747Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T09:36:38.7630581Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T09:36:38.7672761Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T09:36:38.7711782Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T09:36:38.7753498Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T09:36:38.7815144Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-09-07T09:36:38.7830797Z http.https://github.com/.extraheader 2025-09-07T09:36:38.7837614Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-09-07T09:36:38.7860434Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-09-07T09:36:38.8109201Z Entering 'android/libs/fbjni' 2025-09-07T09:36:38.8136624Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8159134Z Entering 'third_party/FP16' 2025-09-07T09:36:38.8189743Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8216324Z Entering 'third_party/FXdiv' 2025-09-07T09:36:38.8245991Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8271715Z Entering 'third_party/NNPACK' 2025-09-07T09:36:38.8300626Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8326900Z Entering 'third_party/NVTX' 2025-09-07T09:36:38.8358299Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8386043Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T09:36:38.8415165Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8440231Z Entering 'third_party/XNNPACK' 2025-09-07T09:36:38.8468702Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8509966Z Entering 'third_party/aiter' 2025-09-07T09:36:38.8541485Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8569793Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T09:36:38.8597340Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8629186Z Entering 'third_party/benchmark' 2025-09-07T09:36:38.8658210Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8683782Z Entering 'third_party/composable_kernel' 2025-09-07T09:36:38.8713434Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8746487Z Entering 'third_party/cpp-httplib' 2025-09-07T09:36:38.8777302Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8804529Z Entering 'third_party/cpuinfo' 2025-09-07T09:36:38.8833940Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8861293Z Entering 'third_party/cudnn_frontend' 2025-09-07T09:36:38.8889970Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8916266Z Entering 'third_party/cutlass' 2025-09-07T09:36:38.8944648Z http.https://github.com/.extraheader 2025-09-07T09:36:38.8976888Z Entering 'third_party/fbgemm' 2025-09-07T09:36:38.9008029Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9037950Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T09:36:38.9067032Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9091582Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T09:36:38.9119137Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9150837Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T09:36:38.9181786Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9207439Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T09:36:38.9238704Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9275410Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T09:36:38.9305844Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9330816Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T09:36:38.9359311Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9384875Z Entering 'third_party/fbgemm/external/json' 2025-09-07T09:36:38.9413106Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9441061Z Entering 'third_party/flash-attention' 2025-09-07T09:36:38.9473273Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9502262Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T09:36:38.9533491Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9564265Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T09:36:38.9592392Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9625202Z Entering 'third_party/flatbuffers' 2025-09-07T09:36:38.9654782Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9685428Z Entering 'third_party/fmt' 2025-09-07T09:36:38.9715085Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9743663Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T09:36:38.9774681Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9801214Z Entering 'third_party/gloo' 2025-09-07T09:36:38.9830112Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9855551Z Entering 'third_party/googletest' 2025-09-07T09:36:38.9885830Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9910456Z Entering 'third_party/ideep' 2025-09-07T09:36:38.9941911Z http.https://github.com/.extraheader 2025-09-07T09:36:38.9968090Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T09:36:38.9995418Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0023828Z Entering 'third_party/ittapi' 2025-09-07T09:36:39.0055175Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0082169Z Entering 'third_party/kineto' 2025-09-07T09:36:39.0111835Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0138398Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T09:36:39.0166762Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0191246Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T09:36:39.0218708Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0243900Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T09:36:39.0272583Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0298568Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T09:36:39.0328144Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0355198Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T09:36:39.0384072Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0409810Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T09:36:39.0440103Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0465689Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T09:36:39.0494080Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0523491Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T09:36:39.0550380Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0577571Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T09:36:39.0607709Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0634328Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T09:36:39.0664047Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0692382Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T09:36:39.0721681Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0747812Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T09:36:39.0775830Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0805616Z Entering 'third_party/kleidiai' 2025-09-07T09:36:39.0836173Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0861468Z Entering 'third_party/mimalloc' 2025-09-07T09:36:39.0891881Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0916019Z Entering 'third_party/nlohmann' 2025-09-07T09:36:39.0942909Z http.https://github.com/.extraheader 2025-09-07T09:36:39.0968753Z Entering 'third_party/onnx' 2025-09-07T09:36:39.0997926Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1037924Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T09:36:39.1070446Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1098469Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T09:36:39.1129911Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1157036Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T09:36:39.1188714Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1214197Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T09:36:39.1245095Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1273113Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T09:36:39.1298620Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1321423Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T09:36:39.1349903Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1379405Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T09:36:39.1408570Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1432410Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T09:36:39.1460182Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1487335Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T09:36:39.1517644Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1544683Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T09:36:39.1576668Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1605051Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T09:36:39.1633090Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1662229Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T09:36:39.1692132Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1733318Z Entering 'third_party/pocketfft' 2025-09-07T09:36:39.1762873Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1788996Z Entering 'third_party/protobuf' 2025-09-07T09:36:39.1819592Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1847881Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T09:36:39.1875900Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1900670Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T09:36:39.1932519Z http.https://github.com/.extraheader 2025-09-07T09:36:39.1962144Z Entering 'third_party/psimd' 2025-09-07T09:36:39.1990963Z http.https://github.com/.extraheader 2025-09-07T09:36:39.2019259Z Entering 'third_party/pthreadpool' 2025-09-07T09:36:39.2048185Z http.https://github.com/.extraheader 2025-09-07T09:36:39.2074655Z Entering 'third_party/pybind11' 2025-09-07T09:36:39.2102765Z http.https://github.com/.extraheader 2025-09-07T09:36:39.2127534Z Entering 'third_party/python-peachpy' 2025-09-07T09:36:39.2155549Z http.https://github.com/.extraheader 2025-09-07T09:36:39.2183300Z Entering 'third_party/sleef' 2025-09-07T09:36:39.2215063Z http.https://github.com/.extraheader 2025-09-07T09:36:39.2242834Z Entering 'third_party/tensorpipe' 2025-09-07T09:36:39.2275490Z http.https://github.com/.extraheader 2025-09-07T09:36:39.2303582Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T09:36:39.2333102Z http.https://github.com/.extraheader 2025-09-07T09:36:39.2360689Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T09:36:39.2389428Z http.https://github.com/.extraheader 2025-09-07T09:36:39.2413617Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T09:36:39.2440865Z http.https://github.com/.extraheader 2025-09-07T09:36:39.2467741Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T09:36:39.2497377Z http.https://github.com/.extraheader 2025-09-07T09:36:39.2525269Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T09:36:39.2555235Z http.https://github.com/.extraheader 2025-09-07T09:36:39.2652336Z A job completed hook has been configured by the self-hosted runner administrator 2025-09-07T09:36:39.2663528Z ##[group]Run '/home/ec2-user/runner-scripts/after_job.sh' 2025-09-07T09:36:39.2666669Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:36:39.2666914Z ##[endgroup] 2025-09-07T09:36:39.2740726Z [!ALERT!] Swap in detected! [!ALERT!] 2025-09-07T09:36:46.5524235Z [!ALERT!] Swap out detected [!ALERT!] 2025-09-07T09:36:58.8688503Z Cleaning up orphan processes